Next Article in Journal
Analysis of Bidirectional ADR-Enabled Class B LoRaWAN Networks in Industrial Scenarios
Next Article in Special Issue
Hybrid Encoding Scheme for AMBTC Compressed Images Using Ternary Representation Technique
Previous Article in Journal
Constitutive Analysis of the Anisotropic Flow Behavior of Commercially Pure Titanium
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Block Compressive Sensing Single-View Video Reconstruction Using Joint Decoding Framework for Low Power Real Time Applications

1
Faculty of Engineering, Science & Technology, Iqra University, Karachi 75500, Pakistan
2
Center for Intelligent Signal and Imaging Research (CISIR), Electrical and Electronics Engineering Department, Universiti Teknologi PETRONAS, Bandar Seri Iskandar 32610, Malaysia
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(22), 7963; https://doi.org/10.3390/app10227963
Submission received: 29 September 2020 / Revised: 25 October 2020 / Accepted: 29 October 2020 / Published: 10 November 2020
(This article belongs to the Special Issue Advances in Signal, Image and Video Processing)

Abstract

:
Several real-time visual monitoring applications such as surveillance, mental state monitoring, driver drowsiness and patient care, require equipping high-quality cameras with wireless sensors to form visual sensors and this creates an enormous amount of data that has to be managed and transmitted at the sensor node. Moreover, as the sensor nodes are battery-operated, power utilization is one of the key concerns that must be considered. One solution to this issue is to reduce the amount of data that has to be transmitted using specific compression techniques. The conventional compression standards are based on complex encoders (which require high processing power) and simple decoders and thus are not pertinent for battery-operated applications, i.e., VSN (primitive hardware). In contrast, compressive sensing (CS) a distributive source coding mechanism, has transformed the standard coding mechanism and is based on the idea of a simple encoder (i.e., transmitting fewer data-low processing requirements) and a complex decoder and is considered a better option for VSN applications. In this paper, a CS-based joint decoding (JD) framework using frame prediction (using keyframes) and residual reconstruction for single-view video is proposed. The idea is to exploit the redundancies present in the key and non-key frames to produce side information to refine the non-key frames’ quality. The proposed method consists of two main steps: frame prediction and residual reconstruction. The final reconstruction is performed by adding a residual frame with the predicted frame. The proposed scheme was validated on various arrangements. The association among correlated frames and compression performance is also analyzed. Various arrangements of the frames have been studied to select the one that produces better results. The comprehensive experimental analysis proves that the proposed JD method performs notably better than the independent block compressive sensing scheme at different subrates for various video sequences with low, moderate and high motion contents. Also, the proposed scheme outperforms the conventional CS video reconstruction schemes at lower subrates. Further, the proposed scheme was quantized and compared with conventional video codecs (DISCOVER, H-263, H264) at various bitrates to evaluate its efficiency (rate-distortion, encoding, decoding).

1. Introduction

The transformation of wireless sensor networks (WSNs) into visual sensor networks (VSNs) has set new horizons in the IT ecological space, making it broadly adapted in various applications. Equally, the coupling of high-quality cameras with sensors has greatly increased the size of the data that has to be managed and transmitted (increasing the computational burden). Furthermore, as the sensor nodes are battery-operated, power utilization is a key concern that must be considered. One potential solution to the aforementioned issues is to reduce the amount of data transmitted using specific image/video compression techniques. In a conventional video capturing arrangement, a compression algorithm is usually applied to the video data once the complete frame set is obtained to exploit the inter-view and intra-view correlations between the frames. Various approaches have been proposed to exploit these correlations for standard video compression [1,2]. The standard complex-encoder simple-decoder method is mostly used by traditional video compression schemes i.e., at encoder heavy processing compared to the decoder. The encoder usually uses the motion estimation and compensation (ME/MC) method to utilize the correlations between the frames that require high computation. Thus, the conventional schemes are not apt for low power real-time applications, i.e., VSN (battery-powered nodes with primitive hardware). Most of the conventional schemes entail the use of extra computational resources for data processing leading to additional power requirements, with few trade-offs between the reconstruction quality and computational constraints [1].
The emergence of compressive sensing (CS) [3] has widened the approach for 3D reconstruction, multi-camera imaging, armed personnel tracing and surveillance, magnetic resonance imaging (MRI) and seismic identification for applications such as surveillance, mental state monitoring, driver drowsiness and patient care, etc. CS is specifically useful for the transmission of correlated data for low power real-time applications, i.e., visual sensors networks and wireless sensors networks. CS has transformed the idea of conventional data compression schemes as it is based on the principle of exemplifying a signal using a much smaller sampling rate than the standard Nyquist rate. Hence, using CS is one of the best solutions. Unlike the conventional schemes it is based on a simple encoder (i.e., transmitting fewer data) and a complex decoder. It is also known that the primary motivation for considering CS is that it offers the potential to reduce sensor costs and computational complexity significantly. For example, it is proposed that the famous single-pixel camera, with only one photo-sensor, could provide an imaging sensor that is much cheaper to build than one with several million photosensors. Moreover, CS is a form of dimensionality reduction only, and it is not specifically a type of compression [4].
CS makes sense only when the dimensionality reduction that it provides takes place directly within the sensing device’s hardware. In other words, the image never exists anywhere in the sensor in its full dimensionality, unlike most of the conventional video coding schemes in which the acquired image is used in its full dimensionality, increasing the load on the sensor node. However, CS can be used for compression, provided that the dimensionality reduction it offers is coupled with quantization and entropy coding to create a bit of data from the CS measurements.
In this paper, a joint decoding (JD) framework is proposed. The proposed scheme aims to decrease the redundancies among video frames at the decoder using side information. The proposed scheme first decodes the encoded keyframes and non-key frames of a video sequence received at the host workstation. After independently decoding the frames, a side information process is initiated. The side information produced is utilized to enhance the reconstruction quality of the non-key frames. The side information process consists of registration and fusion steps used to exploit the inter/intra view correlation between the frames to generate a frame prediction. Next, the difference among the predicted frame measurements and the acquired measurements is calculated to minimize the prediction errors. The difference at the level of the measurement, known as the residual measurements, is recovered to produce the residual. The final frame reconstruction is performed by adding the residual and the predicted frame to compensate for the difference.
The rest of the paper is organized as follows: Section 2 presents an overview of the various existing CS-based video reconstruction coding schemes. A detailed explanation of the proposed JD scheme for single view video reconstruction using block CS is provided in Section 3. In Section 4 the experimental results are presented and the conclusions are drawn in Section 5.

2. Literature Review

Compressive sensing (CS) has emerged as one of the most substantial mechanisms for compressing visual data (video) in recent times. The principle is to sample each video frame independently using CS (encoder), and then at a decoder the correlations within the video frames are exploited using joint reconstruction schemes for image reconstruction. In literature various video reconstruction coding schemes have been proposed (dictionary-based coding, 3D-transformation coding, residual-based coding, etc), each having their benefits and challenges. In the following, only the state-of-the-art residual coding schemes are discussed as the proposed model relies on residual-based coding and will be compared with these schemes. However, complete explanations of the other coding schemes can be found in [5,6,7,8,9,10,11].
In [12], the modified-CS-residual scheme is presented. The scheme is based on a residual reconstruction approach in which side information helps manage the reconstruction issue related to sparse signals (minutest quantity of linear projections). The least mean squares or Kalman filter estimation methods are used to produce side information. However, this side information can be exposed to errors. The proposed scheme aims to counter the convex relaxation problem correlated to data restraints and sparsity outside the side information.
A reconstruction scheme known as k-t FOCUSS is presented in [13,14]. It is also built on a similar residual reconstruction idea. The scheme first generates side information assuming n keyframes using disparity (compensation/estimation) predictions and a residual encoding approach. This leads to an ideal sample allocation among the estimated and residual steps. The residuals among the bidirectional (DC/DE) estimation of each keyframes and non-key frame is used to attain the reconstruction.
A DC/DE-based reconstruction scheme with residuals is introduced in [15]. The scheme generates the side information by integrating DC/DE-based prediction and is referred to as DC-BCS-SPL. The final reconstruction is aided by calculating the residuals among the side information and the original view.
Also, in [16], a DC/DE-based joint reconstruction scheme is proposed. The proximal gradient method is adopted by the proposed scheme to resolve the optimization problem.
The motion (compensation/estimation)-based scheme is introduced in [17]. The proposed MC/ME scheme is incorporated into the block compressive sensing (BCS-SPL) video restoration method and called MC-BCS-SPL. In this approach, each frame of the video sequence is initially sampled using random block-based CS measurements (at the encoder) and transmitted to the decoder. The received encoded measurements are then decoded and the proposed residual dependent ME/MC approach is applied to generate the final view. The proposed scheme generates video frames alternately along with the motion fields associated with the frames i.e., using one to improve the other iteratively.
In [18], a joint reconstruction algorithm using MC/ME and fusion is presented. The proposed scheme uses the down-sample approach (lowering sampling rate of the views) and then MC/ME and a fusion approach are applied on the down-sampled views to produce a view prediction, which helps generate the final view.
A view estimation and residual reconstruction-based joint reconstruction method is introduced in [19]. The method integrates a MC/ME approach into the reconstruction to generate the side information. The side information then helps in the reconstruction of the final output.
The residual coding schemes discussed above help improve the reconstruction quality of the video frames, however, they are exposed to a few issues such as imprecise estimation or computational intricacy. For example, Ref. [12] claims that the proposed scheme is more suitable for video applications and uses least mean squares (LMS) or Kalman filter prediction methods. However, the scheme is exposed to prediction errors. The schemes proposed in [13,14,15,16,17,18,19] are mostly based on DC/MC and DE/ME prediction methods. In such a scheme, precise predictions are hard to achieve if basic transformation (translation/affine) models are used. This is because video captured from different view angles might display some distortions that are hard to handle by using fundamental transformations. Thus, the proposed scheme usually uses a more complex transformation model to resolved prediction issues. However, such a solution might lead to an additional computational burden. A summary of CS-based residual coding schemes for single view video reconstruction is presented in Table 1.

3. Proposed Joint Decoding (JD) Framework for Video

In this section, the proposed joint decoding (JD) framework for single view video is discussed. The archetype of the complete system is shown in Figure 1. At the encoder side, we consider a visual node S, capturing a scene (video), and block-based compressive sensing (BCS) is applied on each frame in the video sequence for encoding. The encoded frames are then independently transmitted to the host workstation.
At the host workstation (decoder), a sequence of J recurrent frames is received from the visual node S, referred to as a group of pictures (GoP). As the video is uninterrupted, it is established that the current GoP is tailed by another GoP. The GoP comprises a keyframe and J-1 non-key frames, represented as FK (the primary) and FNK. The encoding of FK and FNK is performed at the subrate of MK and MNK, respectively, with MK > MNK.
At the visual node (encoder), each frame Fx (where x is the frame number) in the video sequence is initially portioned into a small block of 16 × 16. Then, the sensing matrix Φx as in Equation (1) is used to sample each block within a frame producing a set of measurements (Yx) as in Equation (2) [20,21,22]:
Φ x = [ Φ x 0 0 Φ x ]
Yx = ΦxFx
The set of measurements (Yx) produced at the encoder is sent to the host workstation (server) independently. At the server side (decoder), firstly, the received encoded frames (Yx) are decoded independently (frame by frame) using the total variation (TV) minimization [23,24] problem till a complete GoP is obtained. The TV minimization makes use of piece-wise smooth features of the signals to provide a better solution within the possible space rather than finding the sparse solution within the transformation domain Ψ. The basic TV minimization function is given in Equation (4):
TV ( F ) = i , j | F i + 1 , j F i , j | + | F i , j + 1 F i , j |
F = argmin F | y Θ F | 2 + λ   TV ( F )
However, the CS reconstruction using TV minimization as in Equation (4) is exposed to extra computational problems that limit its use for CS, i.e., it is harder to computationally access and explain certain properties of TV minimization (non-differentiable, non-linear) than ℓ1 minimization. To counter such issues, a scheme [24] referred as TV-AL3 is presented that uses the conventional augmented Lagrangian (AL) with variable splitting and alternating direction method. The approach decreases the computational burden and provides the same output as standard TV.
After obtaining the complete GoP, the proposed joint decoding (JD) method is implemented to utilize the inter and intra-view correlations among the frames. As shown in Figure 1, the first frames of the current and next GoP’s aid as the keyframes FK (reference frames) for the JD to produce side information of frames (prediction, residual) that will help in enhancing the quality of (J-1) non-key frames FNK of the current GoP.

3.1. Correlation Estimation of the CS Measurements among Adjacent Frames

A correlation estimation of the CS measurements among the adjacent frames is presented in this section. As discussed earlier, adjacent frames with a video sequence have a high inter-frame level correlation with each other. Consequently, it can also be assumed that the CS measurements of such frames are also highly correlated. Although, acquiring CS measurements is entirely different from linear transformations, as CS measurements form a random Gaussian distribution. The correlation among the two random entities can be estimated by using Pearson’s correlation coefficient. It is defined as “the correlation coefficient of two random variables is a measure of their linear dependence”. Consider the CS measurements of two adjacent frames Yn and Yn+1 of a video sequence, then their correlation coefficient is estimated as:
ρ ( Y n ,   Y n + 1 ) = 1 φ 1 n 1 φ ( Y n μ Y n σ Y n ) ( Y n + 1 μ Y n + 1 σ Y n + 1 )
where, φ is the number of measurements, μ Y n and μ Y n + 1 are the mean of Yn and Yn+1, respectively, while σ Y n and σ Y n + 1 are the standard deviation of Yn, and Yn+1, respectively.
The evaluation of CS measurement correlation among various frames is performed on various standard grayscale CIF video sequences (Foreman, Coastguard, Container, Hall Monitor, Mother-Daughter). Each video sequence contains 300 frames and is divided into GoPs of size 8. The CS measurements for each frame within a video sequence are obtained at a measurement rate of 0.5. The CS measurement correlation coefficient for each non-keyframe for various video sequences is estimated and shown in Figure 2. The results presented are an average of all the correlations for non-key frame w.r.t keyframes. It can be noticed that all the frames within the video sequence show a high correlation, i.e., above 0.92. Moreover, for moderate motion content video sequences, the correlation is also higher.

3.2. Joint Decoding Framework

This subsection discusses in detail the decoding process of the proposed JD scheme. The proposed JD consists of three significant steps, as presented in Figure 3.
The descriptions of each step are provided in the following sections.

3.2.1. Frame Estimation

This step proposes a frame estimation technique using registration and fusion methods. The idea is to utilize the frames’ correlations to estimate the J-1 non-key frame (F′NK) in the GoP from the keyframes (F′K). The proposed method assists in utilizing the inter and intra-view correlations between the frames and produce a set of predicted non-key frames.

Registration Approach

The frame estimation step is initialized by performing registration (intensity) on the two independently generated keyframes F′K in the GoP. The aim is to align the F′K on the same plane as of F′NK (F′K is aligned to F′NK, and the redundancies between them are exploited). Firstly, a phase correlation method (finding the gross alignment) is used on the F′K and F′NK frames to calculate the initial transformation matrix. Next, the translation transformation over affine is used for aligning F′K w.r.t. F′NK to produce transformed F′K called F″KT. The reason for using a translation transform is that affine transform makes sense when multiple images are not on the same plane and are to be rectified, whereas, in the case of a video sequence, each frame is usually on the same plane, so it would be sufficient to consider translation transform.
The similarity metric and optimization function is then applied on the transformed frame F″KT to estimate the registration precision and generate the final registered image F″K, as shown in Equation (6):
F K = argmin N | | y L F KT x L F NK P F KT F NK ( y , x ; a ) l o g 2 ( P F KT F NK ( y , x ; a ) P F KT ( x ) P F NK ( y ; a ) ) | |
The similarity metric makes use of the mutual information (MI) method where the optimizer uses one evolutionary (OE). The evolutionary optimizer is used to exploits the alike alignment that usually occurs within the frames of a video sequence that better helps the OE optimizer than the other optimizers such as gradient descent (GD). The optimizer is one of the vital parameters of the registration process that defines the method to exploit the attained similarity metric M and produce a final result. Once both the keyframes F″K is registered, a wavelet-based fusion process is applied on them to produce the predicted frame FP.

Fusion Approach

In the fusion process, the registered keyframes F″K are down-sampled up to three levels using symlet 4-tap filter into their respective approximation (A) and detail (D) coefficient maps. Then, the point-to-point operations are performed to fuse the A and D coefficients in the two decomposition maps. The coefficients A and D are set as shown in Equations (7) and (8), respectively:
AMean(a,b) = (AK (a,b) + ANK (a,b))/2
DMean(a,b) = (DK (a,b) + DNK (a,b))/2
The average magnitudes for each A and D coefficient having a similar coordinate of the two decomposition maps is calculated. The average value obtained then serves as the output for the fused map (FM). Once the fused map is generated, an inverse transformation is applied to generate the FP’s predicted frames. The proposed frame estimation method predicts the object motions and produces predicted frames FP:
Fp = Transform−1(FM)
The defined set of parameters for registration and fusion processes remained best when related to other sets of parameters for various video sequences that have been used.

3.2.2. Residual Reconstruction

In this step, the predicted frames FP is projected onto the measurement basis YP = Φx IP. Once, the predicted measurements YP is generated, they are then differentiated from the given frame measurements Yx to produce the residual measurement Yr as in Equation (10):
Yr = Yx − YP
The obtained residual measurements Yr are then decoded by using BCS-TV-AL3 method to generate the residual frames Fr.

3.2.3. Final Frame Reconstruction

The final reconstructed frames F″NK within the GoP are generated by adding the Fr and FP frames. It is a standard point-to-point addition that is expressed in Equation (11). By doing so, uniformity in terms of frame measurements (Y) is achieved, i.e., the measurements computed for F″NK is to some extent equal to the measurements YNK:
F″NK = Fr + FP
After the keyframes, F′K (F0 and FJ from Y0 and YJ) are reconstructed using BCS-TV-AL3, they are used as the reference frames for reconstructing the non-key frames F′NK between them.
The proposed scheme produces the non-key frame F″1 from Y1, F0, and FJ; similarly F″2 is generated from Y2, F0 and FJ. The method endures for all the remaining non-key frames. The reconstructed frames’ quality is expected to drop as the distance of the key frames from the non-key frames increases. Thus, the quality of the reconstruction may decrease with the increase in the GoP size (J). A complete GoP (J = 8) reconstruction for News video sequence is shown in Figure 4. The highlighted red dotted circles clearly show that as the difference among the key frames and non-key frame increases the reconstruction quality decreases.

4. Experimental Results

4.1. Setup

In this subsection, the validation of the proposed JD scheme incorporated with TV-AL3 (i.e., JD-TV) in terms of performance and efficiency is presented. A standard set of grayscale CIF [25] and HD resolution [26] video sequences with sizes of 352 × 288 and 1024 × 678 are selected on which the proposed scheme is applied for evaluation. The selected video sequences are based on different parameters such as # of frames, motion content type (slow, moderate, fast). The details of the video sequences (# frames, motion type) are presented in Table 2. The video sequence indicated as low, moderate, and high contents are based on spatial details plus camera and object movement.
The experimental setup comprises of the application of proposed JD-TV having various GoP (J) sizes (3, 5, and 8) to validate its efficiency at various levels. The assessment is performed by observing the peak signal to noise ratio (PSNR) at various sampling rates (subrate). Subrate or sampling rate is the number of recorded samples (pixel) per unit distance when converting from an analog signal to digital and is give as S = MB/N. Where, M is the number of CS measurements, N = B × B (B = block size).
Also, the structural similarity index (SSIM) is also recorded as it is anticipated as more precise and consistent (human visual perception) than PSNR. All PSNR and SSIM values presented are the mean of five independent experiments. The values are averaged because the Φ is of random nature, and thus, the image quality may vary. A smaller block size of 16 × 16 is implemented compared to larger ones (32 × 32, 64 × 64), as it is observed that larger block size will produce more samples and will require additional time and energy to transmit. Although, the larger block size provides better reconstruction quality but at the expense of complexity. The selection of block size is a tradeoff between reconstruction quality and computational complexity. In this regard, it is not feasible for a battery-powered device to always encode and transmit the captured images at larger block size. In addition, most of the research works focused on low powered application have adopted block size of 16 × 16 or 32 × 32 as it provides better image quality with less computational complexity. The encoding of all keyframes are performed at a fixed subrate of 0.5 while the non-key frames within a GoP are encoded at various lower subrates (0.05, 0.1, 0.15, 0.2, 0.25, 0.3).

4.2. Effect of GoP Size on the Performance of Proposed JD-TV

In this subsection, the proposed JD-TV is assessed and equated with independent BCS-TV-AL3 using various GoP sizes. Table 3 shows the impact of three different GoP sizes (3, 5, 8) on the performance of the proposed JD at various subrates for several video sequences.
The results shown are the average values of the non-key frames obtained for each complete video sequence. The result shows that the proposed scheme performs well when compared with independent BCS-TV-AL3. For low, moderate, and high content types, the improvement over independent BCS-TV-AL3 is 3 dB- 7 dB, 2.5 dB-5 dB and 2 dB-4 dB, respectively, on average for all GoP sizes. Also, it is also noticed that the gain for reconstructed low content video sequence is better than moderate and high content as the association between the frames is higher than in moderate and high content videos. The better correlation results in more precise frame estimation and residual reconstruction.
It is also observed that the larger the GoP size, the lower the PSNR gain will be. This is because the principle of the proposed scheme is to reconstruct the non-key frames by utilizing the correlation of the keyframes. Thus, the non-key frames closer to the keyframe are more correlated than those further away. For GoP = 3 the average gain is 3dB to ~6dB, while for GoP = 8 ~2dB to 4dB is achieved.
Additionally, as the subrate increases, the gain decreases. In other words, as the keyframes, FK is encoded at a subrate higher than that of non-key frames FNK, it produces a larger measurement, covering the related smaller set of FNK measurements. This, in result decreases the estimation errors of FNK that arises due to smaller set measurements and generates a better form of FNK.
The reconstructed frames’ visual quality is also tested by using SSIM metric on the same video sequences used for PSNR. The SSIM graphs of various video sequences with GoP sizes of 3, 5, and 8 at different subrates are presented in Figure 5.
The SSIM shows a similar trend in terms of gain as of PSNR. Moreover, the result also indicates an enhancement in visual quality and a notable increase over the independent scheme. Also, the proposed scheme is tried at higher GoP sizes (16, 32). However, the frame reconstruction was not substantial enough to be imitated. In the case of VSN, the frequent transmission of the keyframes is not encouraged due to limited battery life as in the instance with GoP size 3 and 5; it will upsurge the computational load at the encoder (Low power node). Thus, GoP size 8 is measured as a more stable point between all the GoP sizes and will later opt-in trials.

4.3. Subrate

This subsection presents the impact of different FK subrate used in the proposed JD-TV on the FNK reconstruction. A GoP size of 8, as indicated from the former experimentation (balance point) instead of 16/32 is opted. The evaluation is carried on two arrangements.
  • First setup, encoding of both FK and FNK at the similar subrate i.e., MK = MNK (0.05, 0.1, 0.15, 0.2, 0.25, 0.3).
  • Second setup, the encoding of FK at a static subrate of MR = 0.5 and FNK at various subrates i.e., MNK (0.05, 0.1, 0.15, 0.2, 0.25, 0.3) with 0.05 interval among the subrate.
The evaluation result shown in Table 4 indicates the importance of subrate on the reconstruction quality of the frames. In other words, FK subrate has a substantial effect on the FNK reconstruction quality. In the first setup, where subrate MK = MNK the reconstruction quality, when compared with traditional scheme, improves on average from 2 dB–3.5 dB. However, when the subrate is MK > MNK, it is noted that the gain increases significantly (3.5 dB–5 dB) when compared with JD-TV (MK = MNK) and BCS-TV. The reason that the reconstruction quality at subrate MK > MNK is better than at subrate MK = MNK is because, when encoding of both FK and FNK is performed at the same subrate, the generated FK does not comprehend enough material that could considerably assist the reconstruction of FNK.

4.4. Proposed JD Framework and HD Video Sequences

In this section, the proposed JD framework is tested with HD video sequences in order to evaluate its effectiveness at various video resolutions. The HD video sequence selected contains both low content and moderate content types. A block size of 16 × 16 and GoP size of 8 is adopted.
The PSNR and SSIM results presented in Table 5 and Table 6 respectively, clearly show that the proposed JD framework also performs well for HD video sequences compared with conventional BCS framework. A similar trend as that of CIF video sequence can be noticeable using JD reconstructions. For low motion video (Lovebird) the gain in term of PSNR and SSIM is more significant than moderate motion videos (Newspaper, Book arrival).

4.5. Visual Result Comparison

Visual quality analysis is one of the vital evaluation parameters for the reconstruction of compressed images. In this subsection, the visual quality of the reconstructed frames using proposed JD is analyzed. The proposed framework performs better in terms of PSNR and SSIM at lower sub-rate. However, it is also important to verify that the frame produced at lower subrate are visually perceptible. Thus, we perform the analysis on three different video sequences having low, moderate and high motion contents.
Figure 6 presents the visual results of the random center frame of each GoP selected from the reconstructed video sequence (the similar visual quality was observed for all center frames of each GoP) by using JD-TV and BCS-TV-AL3. The result shows that the visual quality of reconstructed frames by using proposed method at lower subrates is better visually perceptible than the conventional method. In addition, for low motion content video sequence (Mother-Daughter) the proposed JD-TV shows higher visual quality (due to precise frame prediction and residual reconstruction) as compared to moderate (News) and higher Motion content video sequences.
It should also be noted that at a lower subrate (0.05) due to the insufficient estimation information of motion few distortions are detected (red dotted circle). Further, for video sequences with fast-moving objects (Mobile Calendar), the JD-TV is vulnerable to a few distortions as emphasized by the red spotted circle.

4.6. Proposed JD-TV v/s Standard CS Video Compression Schemes

In order to analyze the performance of the proposed JD-TV methodically it is compared with conventional CS-based schemes i.e., MS-residual [12], k-t FOCUSS [13], MC-BCS-SPL [17] as referred in Section 2. The results presented in Table 7 (gain of proposed and conventional schemes w.r.t independent BCS) are the average of the total frames obtained at a block size of 16 × 16 and GoP of 8.
The results in Table 7 clearly show that the proposed JD provides considerable gain for various motion content video sequences at lower subrates than conventional CS-based schemes. Further, is can also be observed that the gain decreases as the subrate increases, which is due to impact of FK subrate as discussed earlier. In addition, the focus of the proposed scheme is to provide better reconstruction at lower subrates.

4.7. Comparison of Proposed JD-TV with Conventional Video Compression Schemes

As discussed earlier, CS is a type of dimensionality reduction such that the signal at no time exists in the sensor in its full dimensionality. CS is based on a simple-encoder complex-decoder archetype, which is in contrast to the conventional video codecs. In order to use CS as a compression mechanism, quantization and entropy coding must be coupled with CS to generate bit from the CS measurements.
In this section, the proposed JD-TV is coupled scalar quantization- adaptive differential pulse code modulation (SQ-ADPCM) referred as SQ-ADPCM-JD to compare it with state-of-the-art video codecs i.e., DISCOVER [27], H.264 (intra, (I-P-P)) [28] and H.263 (intra, (I-P-P)) [29]. The results shown in Figure 7 are the average values of the frames obtained for each complete video sequence at group of picture (GoP) = 3 and block size = 16 for all implementations.
It can be noted from the results that the CS codec, coupled with SQ-ADPCM, performs better than the standard video codec at various bitrates. For example, when compared with H.263 (Intra) and H.264 (Intra), the CS codec outperforms them for all video sequences at various bitrates. For the case of moderate content type video (Foreman, Coastguard), the CS codec performs noteworthily better than H.263 (I-P-P) at various bitrates, while it should also be noted that the performance of the proposed SQ-ADPCM incorporated with the JD scheme performs better than DISCOVER and H.264 (I-P-P) at lower bitrates. As mentioned earlier, the proposed JD-TV make use of keyframes that contains larger set of measurements (higher subrate) to improve the non-key frames with lower measurement set (low subrates). Thus, the key frames having larger set of measurements superimposes the correlated smaller measurements set encompass by non-key frames. This, helps to decrease the non-key frame prediction errors, that arises due to smaller measurements set and predict a better version of non-keyframes. Further, DISCOVER and H.264 (I-P-P) schemes use the feedback channel to improve the reconstruction of the keyframes.

4.8. Number of Bits

The proposed JD scheme is developed with the idea of benefiting low power real-time applications (VSNs) in terms of efficiency, data transmission and power utilization. This subsection observed the bit rate savings among the independent BCS-TV-AL3 and the proposed JD-TV for different video sequences at the various reconstruction qualities (PSNR).
The results presented in Table 8 shows the reconstruction quality at different numbers of bits for different video sequences. The results obtained can also be calculated as given in [30]. It can be noted that the number of bits essential for conventional BCS and the proposed JD schemes to achieve the same PSNR varies. The proposed scheme utilizes minimum number of bits as compared to the independent BCS scheme to achieve the same PSNR i.e., for higher motion content video sequence the average saving rate is ~42%, while for lower content video sequence the average saving rate is ~65% for all videos. This is due to the fact that the proposed JD offers improved reconstruction quality at lower bit rates. As the proposed scheme provides better reconstruction at lower bits this helps to reduce the computational burden from the encoder as the encoder will be transmitting small amounts of data. On average, the number of bits saved by the proposed scheme contrary to the independent scheme is ~50% at different reconstruction qualities.

4.9. Execution Time (Encoding/Decoding Complexity)

In this subsection, the encoding and decoding complexity (average compression and reconstruction time) of the proposed JD-TV and other conventional video schemes i.e., DISCOVER [27], and H.264 (intra) [28] with GoP size 3 and block size 16 for various video sequences is presented. Also, the reconstruction complexity of the proposed schemes is compared with other CS-based reconstruction schemes. All the schemes are implemented using MATLAB (R2019a) running on a computer equipped with an Intel® CoreTM i7-6700 CPU @ 3.4 GHz and 16 GB RAM.
The results presented in Table 9 shows that the average encoding time of the proposed JD-TV, DISCOVER [27], and H.264 (intra) [28] ranges from 6.31 s–7.38 s, 31.05 s–60.08 s, and 62.40 s–125.45 s, respectively. It is also observed that at higher subrates, the proposed JD-TV requires more encoding time than lower subrates. In contrast to conventional video codecs, the proposed JD-TV codec takes lesser time to encode the video frames i.e., for all types of video sequence (low, medium, high) the average encoding time of the proposed JD-TV method is significantly lower than the DISCOVER and H.264 (intra) video codecs. For example the encoding time required by JD-TV to encode the Foreman video sequence is 6.94 s which, is 4–5 times less than DISCOVER (36.40 s) and 9–10 times less than the H.264 (Intra) (78.50 s). A similar trend can be observed for other video sequences. Thus, this validates the fact that the conventional video schemes’ encoders are exposed to heavy computation burden and are not suitable for low power real-time applications.
We also observed the decoding complexity (average reconstruction time) of the proposed JD-TV, DISCOVER, and H.264 (intra). The results presented in Table 10 shows that the proposed JD-TV requires more reconstruction time as compared to the conventional codecs, i.e., its decoder has a heavy computational complexity when compared with DISCOVER and H.264 (intra). This validates the fact the CS-based codecs are based on a simple encoder and complex decoder, unlike conventional codecs (DISCOVER and H.264 (intra)). Also, it should also be noted that the results shown are a trade-off between encoding/decoding complexity and rate-distortion performance and has a weak comparability aspect.
Additionally, the average reconstruction time of the proposed JD-TV and other CS reconstruction schemes i.e., MS-residual [12], k-t FOCUSS [13], and MC-BCS-SPL [17], is also shown in Figure 8. Overall, the proposed JD-TV executes about 2–3 times faster than MC-BCS-SPL and kt-Focuss whereas, the proposed JD-TV outperforms MS-residual. This reason is that the proposed scheme incorporates BCS-TV-AL3 (less complex) for initial reconstruction and the simplified process of predicting the non-key frames.

5. Conclusions

In this paper, a joint decoding (JD) framework is proposed that reduces the redundancies present between video frames at the decoder and reduces the computational burden on the encoder. The proposed scheme makes use of an efficient registration and fusion method to generate side information that helps in the reconstruction of final frame sequences. The simulation results are based on different arrangements with different subrates and GoP sizes (3, 5, 8). The results show the effect of different GoP sizes on the reconstruction quality i.e., smaller GoP = 3 provides improved reconstruction quality compared to larger ones. The comprehensive experimental analysis proves that the proposed JD-TV performs notably better than the independent BCS-TV-AL3 scheme at different subrates for various video sequences having low, moderate and high motion contents. In addition, the proposed scheme outperforms the conventional CS video reconstruction schemes at lower subrates. Further, when compared with conventional video codecs (DISCOVER, H-263, H264) the proposed framework shows notable performance at various bitrates. Further, the coding efficiency in terms of encoding and decoding time of the proposed and conventional video codes are also compared.
The developed scheme can be used in real-time transport safety applications such as driver warning systems for motorbikes or cars, or driver drowsiness detection. Similarly, it can be applied for real-time visual patient monitoring, MRI, patient care and remote sensing.

Author Contributions

Conceptualization: M.E.; Methodology: M.E.; Software: M.E. and S.H.A.; Validation: M.E., S.H.A., and K.R.; Formal analysis: M.E., S.S.A.A.; Investigation: M.E. and S.H.A.; Resources: M.E. and K.R.; Data curation: M.E. and S.H.A.; Writing—original draft preparation: M.E.; Writing—review and editing: M.E. and S.S.A.A.; Visualization: M.E., K.R. and S.S.A.A.; Supervision: S.H.A. and K.R.; Project administration: M.E. and S.H.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Acknowledgments

The author would like to acknowledge the technical and administrative support of Sunway University, Malaysia, Iqra University, Pakistan and Universiti Teknologi PETRONAS, Malaysia. This research is partially supported by Iqra University, Pakistan and Ministry of Education, Malaysia under the Higher Institute Centre of Excellence (HICoE) awarded to Center for Intelligent Signal and Imaging Research, Universiti Teknologi PETRONAS, Malaysia.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Puri, R.; Majumdar, A.; Ishwar, P.; Ramchandran, K. Distributed Video Coding in Wireless Sensor Networks. IEEE Signal Process. Mag. 2006, 23, 94–106. [Google Scholar] [CrossRef]
  2. Ebrahim, M.; Chong, C.W. A Comprehensive Review of Distributed Coding Algorithms for Visual Sensor Network (VSN). Int. J. Commun. Netw. Inf. Secur. (IJCNIS) 2014, 6, 104–117. [Google Scholar]
  3. Donoho, D.L. Compressed Sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  4. Ebrahim, M.; Chong, C.W.; Adil, S.H.; Raza, K. Block Compressive Sensing (BCS) Based Low Complexity, Energy Efficient Visual Sensor Platform with Joint Multi-Phase Decoder (JMD). Sensors 2019, 19, 2039. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Chen, X.; Frossard, P. Joint reconstruction of compressed multi-view images. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 1005–1008. [Google Scholar]
  6. Li, X.; Wei, Z.; Xiao, L. Compressed sensing joint reconstruction for multi-view images. Electron. Lett. 2010, 46, 1548–1550. [Google Scholar] [CrossRef]
  7. Wakin, M.B. A manifold lifting algorithm for multi-view compressive imaging. In Proceedings of the Picture Coding Symposium, Chicago, IL, USA, 6–8 May 2009. [Google Scholar]
  8. Ebrahim, M.; Chong, C.W. Multi-view Image Block Compressive Sensing with Multi-phase Joint Decoding for Visual Sensor Network. ACM Trans. Multimed. Comput. Commun. Appl. 2015, 12, 23. [Google Scholar]
  9. Park, J.Y.; Wakin, M.B. A geometric approach to multi-view compressive imaging. EURASIP 2012, 37. [Google Scholar] [CrossRef] [Green Version]
  10. Marcia, R.; Willet, R. Compressive coded aperture video reconstruction. In Proceedings of the European Signal Processing Conference (EUSIPCO), Lausanne, Switzerland, 25–29 August 2008. [Google Scholar]
  11. Park, J.Y.; Wakin, M.B. A multiscale framework for compressive sensing of video. In Proceedings of the Picture Coding Symposium, Chicago, IL, USA, 6–8 May 2009; pp. 1–4. [Google Scholar]
  12. Lu, W.; Vaswani, N. Modified compressive sensing for real-time dynamic MR imaging. In Proceedings of the 2009 16th IEEE International Conference on Image Processing (ICIP), Cairo, Egypt, 7–10 November 2009; pp. 3045–3048. [Google Scholar] [CrossRef]
  13. Jung, H.; Sung, K.; Nayak, K.S.; Kim, E.Y.; Ye, J.C. k-t FOCUSS: A general compressed sensing framework for high resolution dynamic MRI. Magn. Reson. Med. 2009, 61, 103–116. [Google Scholar] [CrossRef] [PubMed]
  14. Jung, H.; Ye, J.C. Motion estimated and compensated compressed sensing dynamic magnetic resonance imaging: What we can learn from video compression techniques. Int. J. Imaging Syst. Technol. 2010, 20, 81–98. [Google Scholar] [CrossRef]
  15. Trocan, M.; Maugey, T.; Tramel, E.W.; Fowler, J.E.; Popescu, P. Compressed sensing of multi-view images using disparity compensation. In Proceedings of the 2010 IEEE International Conference on Image Processing, Hong Kong, China, 26–29 September 2010; pp. 3345–3348. [Google Scholar] [CrossRef] [Green Version]
  16. Chang, K.; Qin, T.; Xu, W.; Men, A. A joint reconstruction algorithm for multi-view compressed imaging. In Proceedings of the 2013 IEEE International Symposium on Circuits and Systems (ISCAS), Beijing, China, 19 May 2013; pp. 221–224. [Google Scholar] [CrossRef]
  17. Mun, S.; Fowler, J.E. Residual reconstruction for block-based compressed sensing of video. In Proceedings of the IEEE Data Compression Conference, Snowbird, UT, USA, 29–31 March 2011; pp. 183–192. [Google Scholar]
  18. Cen, N.; Guan, Z.; Melodia, T. Joint decoding of independently encoded compressive multi-view video streams. In Proceedings of the Picture Coding Symposium (PCS), San Jose, CA, USA, 8–11 December 2013. [Google Scholar]
  19. Trocan, M.; Tramel, E.W.; Fowler, J.E.; Pesquet, B. Compressed sensing recovery of multi-view image and video sequences using signal prediction. Multimed. Tools Appl. 2014, 72, 95–121. [Google Scholar] [CrossRef]
  20. Candes, E.; Romberg, J. Sparsity and incoherence in compressive sampling. Inverse Probl. 2007, 23, 969–985. [Google Scholar] [CrossRef] [Green Version]
  21. Gan, L. Block Compressed Sensing of Natural Images. In Proceedings of the 15th International Conference on Digital Signal Processing, Cardiff, UK, 1–4 July 2007; pp. 403–406. [Google Scholar]
  22. Ebrahim, M.; Chong, C.W.; Adil, S.H.; Nawaz, D. A Performance Comparative Analysis of Block Based Compressive Sensing and Line Based Compressive Sensing. Eng. Technol. Appl. Sci. Res. 2018, 8, 2809–2813. [Google Scholar]
  23. Chambolle, A.; Lions, P.L. Image recovery via total variation minimization and related problems. Numer. Math. 1997, 76, 167–188. [Google Scholar] [CrossRef]
  24. Li, C. Compressive Sensing for 3d Data Processing Tasks: Applications, Models and Algorithms. Ph.D. Thesis, Rice University, Houston, TX, USA, 2013. Available online: http://hdl.handle.net/1911/70314 (accessed on 13 March 2020).
  25. YUV Video Test Sequences, Retrieved 15 January 2015. Available online: http://www.codersvoice.com/a/webbase/video/08/152014/130.html (accessed on 21 April 2020).
  26. HD Video Sequence Book Arrival, Newspaper, LoveBirds Multiview Video Sequence—Courtesy IRCCyN IVC DIBR. Available online: http://ivc.univ-nantes.fr/en/databases/DIBR_Videos/ (accessed on 21 April 2020).
  27. Artigas, X.; Ascenso, J.; Dalai, M.; Klomp, S.; Kubasov, D.; Ouaret, M. The DISCOVER codec: Architecture, techniques and evaluation. In Proceedings of the Picture Coding Symposium, Lisbon, Portugal, 7–9 November 2007. [Google Scholar]
  28. International Telecommunication Union (ITU). H.264: Advanced Video Coding for Generic Audiovisual Services. ITU-T Recommendations for h.264, 2005 132-International. Available online: https://www.itu.int/rec/T-REC-H.264 (accessed on 18 October 2020).
  29. Telecommunication Union (ITU). H.263: Video Coding for Low Bit Rate Communication. ITU-T Recommendations for h.263. 2005. Available online: https://www.itu.int/rec/T-REC-H.263/ (accessed on 18 October 2020).
  30. Serge. Bjontegaard Metric Calculation (BD-PSNR). 2020. Available online: https://www.mathworks.com/matlabcentral/fileexchange/41749-bjontegaard-metric-calculation-bd-psnr (accessed on 21 October 2020).
Figure 1. Archetype of the complete system for single-view video coding using the proposed JD (joint decoding).
Figure 1. Archetype of the complete system for single-view video coding using the proposed JD (joint decoding).
Applsci 10 07963 g001
Figure 2. Correlation analysis of CS measurements among the adjacent frames of various CIF video sequences.
Figure 2. Correlation analysis of CS measurements among the adjacent frames of various CIF video sequences.
Applsci 10 07963 g002
Figure 3. Reconstruction of frames using the proposed joint decoding (JD) framework.
Figure 3. Reconstruction of frames using the proposed joint decoding (JD) framework.
Applsci 10 07963 g003
Figure 4. Visual quality analysis for a complete GoP (J = 8) of News video sequence at a subrate of 0.1 using proposed JD-TV framework.
Figure 4. Visual quality analysis for a complete GoP (J = 8) of News video sequence at a subrate of 0.1 using proposed JD-TV framework.
Applsci 10 07963 g004
Figure 5. Average SSIM evaluation at GoP sizes of 3, 5, and 8 for different video sequences at different subrates for independent BCS-TV-AL3and proposed JD.
Figure 5. Average SSIM evaluation at GoP sizes of 3, 5, and 8 for different video sequences at different subrates for independent BCS-TV-AL3and proposed JD.
Applsci 10 07963 g005
Figure 6. Visual quality analysis of different video sequences at multiple subrates using independent BCS-TV-AL3 and the proposed JD-TV.
Figure 6. Visual quality analysis of different video sequences at multiple subrates using independent BCS-TV-AL3 and the proposed JD-TV.
Applsci 10 07963 g006aApplsci 10 07963 g006b
Figure 7. PSNR comparison of proposed SQ-ADPCM-JD with conventional codecs i.e., DISCOVER, H.264 (intra, (I-P-P)) and H.263 (intra, (I-P-P)) at various bitrate for different video sequences.
Figure 7. PSNR comparison of proposed SQ-ADPCM-JD with conventional codecs i.e., DISCOVER, H.264 (intra, (I-P-P)) and H.263 (intra, (I-P-P)) at various bitrate for different video sequences.
Applsci 10 07963 g007
Figure 8. Average execution time (sec) comparison of proposed CS based JD-TV codec with other CS codecs for various video sequences.
Figure 8. Average execution time (sec) comparison of proposed CS based JD-TV codec with other CS codecs for various video sequences.
Applsci 10 07963 g008
Table 1. Summary of various CS-based residual coding schemes for visual reconstruction.
Table 1. Summary of various CS-based residual coding schemes for visual reconstruction.
WorkSchemeDescriptionIssues
[12]Kalman-Filter (KF) based PredictionMake use of side information to handle the sparse signals reconstruction issue (minimum number of linear projections).
It also resolves the convex relaxation issue linked with data restraints and sparsity, outer the side information.
It is usually viable for video applications, as it assumes that the sparsity pattern develops progressively from frame to frame.
Besides, Prediction make use of state dynamics and measurement models. The prediction might fail due to incorrect initialization of the model and filter.
A tree-structured KF algorithm can be used to cater the issue at the cost of higher computational resources.
[13]Disparity Estimation & Compensation
(DE/DC) based Prediction
Incorporates disparity estimation and compensation prediction methods into the reconstruction process of BCS-SPL that generate side information aiding final reconstruction.Introduces discontinuities at the block borders (blocking artefacts).
Accurate predictions are hard to achieve with fundamental estimation, and compensation algorithms as images/frames captured from different view angles may exhibit some deformations.
May result in producing false edges and ringing effects.
Requires more complex estimation and compensation algorithms that result in additional computational burden.
Significant improvements at higher subrate as compared to lower subrate i.e., the scheme does not accurately predict the motions due to the smaller number of measurements at lower subrates that leads to low-quality initial reconstructions
[14]Solve the optimization problem by implementing the proximal-gradient method, and side information is generated by DE/DC prediction approach.
[15,16]Motion Estimation & Compensation
(ME/MC) based Predication
Uses motion-based prediction and residual encoding to optimize the sample allocation among the estimated and residual encoding steps.
[17]The MC/ME approach is incorporated with the BCS-SPL reconstruction process for the video. The video sequence frames are generated alternatively i.e., one helps to improve the quality of the other iteratively.
[18]The scheme uses low sample rate, ME/MC and fusion method to produce a view prediction, which helps in the generation of an ultimate view for the multi-view video.
[19]The MC/ME approach is incorporated with the BCS-SPL reconstruction process for multi-view video.
Table 2. Various grayscale CIF and HD standard video sequences.
Table 2. Various grayscale CIF and HD standard video sequences.
Video SequenceNo. FramesContent-Type
CIF
(352 × 288)
Hall-Monitor300Low
Mother Daughter300Low
Coast-Guard300Moderate
Foreman300Moderate
Mobile Calendar 300High
Stefan300High
HD
(1024 × 678)
Love Birds300Low
News300Moderate
Book-Arrival300Moderate
Table 3. Average PNSR evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different low, moderate, and high motion content video sequences at various GoP.
Table 3. Average PNSR evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different low, moderate, and high motion content video sequences at various GoP.
Low Motion Content VideoSequenceMother-Daughter
Subrate0.050.100.150.200.250.30
BCS-TV24.8830.3632.3533.7335.5537.15
JD-TV GoP336.3540.1041.0341.4141.9442.41
JD-TV GoP534.1635.5536.7338.4339.8840.72
JD-TV GoP832.0034.2635.8137.6538.6339.61
Hall Monitor
Subrate0.050.100.150.200.250.30
BCS-TV21.2023.7125.2626.6828.1829.71
JD-TV GoP332.3132.7033.2833.9334.4535.08
JD-TV GoP529.5231.5632.1032.9833.6434.26
JD-TV GoP828.9830.6531.7732.8133.8334.85
Moderate Motion Content Video SequenceForeman
Subrate0.050.100.150.200.250.30
BCS-TV23.6625.3827.0529.030.7033.83
JD-TV GoP329.8231.9432.3133.5934.1435.81
JD-TV GoP526.329.931.0632.8433.5935.03
JD-TV GoP826.2927.8630.0531.4232.9634.61
Coast Guard
Subrate0.050.100.150.200.250.30
BCS-TV21.8623.9825.0126.1527.1827.99
JD-TV GoP329.1829.7330.2230.931.5132.00
JD-TV GoP525.4326.8427.5128.629.2730.02
JD-TV GoP824.5325.6126.6927.628.3829.14
High Motion Content Video SequenceMobile Calendar
Subrate0.050.100.150.200.250.30
BCS-TV17.7919.7120.6921.7722.7123.61
JD-TV GoP324.5425.3825.8226.4927.227.76
JD-TV GoP522.4623.4924.2624.9525.7126.39
JD-TV GoP821.2622.3923.3524.0824.925.76
Stefan
Subrate0.050.100.150.200.250.30
BCS-TV19.9221.7622.8924.1425.2326.42
JD-TV GoP324.327.0627.1328.7129.5330.44
JD-TV GoP523.6825.3426.5927.8728.5229.76
JD-TV GoP823.3925.0826.2727.0727.9329.00
Note: The bold values represent the highest PSNR (dB) achieved at specific subrate for video sequence.
Table 4. Average PNSR evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different low, moderate and high motion content video sequences at GoP = 8.
Table 4. Average PNSR evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different low, moderate and high motion content video sequences at GoP = 8.
Low Motion Content VideoHall Monitor
Subrate0.050.100.150.200.250.30
BCS-TV-AL321.7023.5324.7626.4428.1629.77
JD-TV (MR = MNR)22.8825.5427.3729.3430.8432.62
JD-TV (MR = 0.5)28.9830.6531.7732.8133.8334.85
Mother Daughter
Subrate0.050.100.150.200.250.30
BCS-TV-AL324.8830.3632.1933.7335.5537.15
JD-TV (MR = MNR)29.0232.1533.8135.5737.8339.15
JD-TV (MR = 0.5)32.0034.2635.8137.6538.6339.61
Moderate Motion Content VideoCoast guard
Subrate0.050.100.150.200.250.30
BCS-TV-AL321.8623.9825.0126.1527.1827.99
JD-TV (MR = MNR)22.3224.9025.3726.4127.6728.58
JD-TV (MR = 0.5)24.5325.6126.6927.628.3829.14
Foreman
Subrate0.050.100.150.200.250.30
BCS-TV-AL323.6625.3827.0529.030.7033.83
JD-TV (MR = MNR)24.9226.3928.1029.8431.3533.35
JD-TV (MR = 0.5)26.2927.8630.0531.4232.9634.61
High Motion Content VideoMobile Calendar
Subrate0.050.100.150.200.250.30
BCS-TV-AL316.6818.5819.5820.5221.5822.50
JD-TV (MR = MNR)17.9919.4320.6521.8022.8824.03
JD-TV (MR = 0.5)20.1521.2822.2422.9723.7924.65
Stefan
Subrate0.050.100.150.200.250.30
BCS-TV-AL319.9221.7622.8924.1425.2326.42
JD-TV (MR = MNR)20.2322.6523.8525.0926.1227.50
JD-TV (MR = 0.5)23.3925.0826.2727.0727.9329.14
Note: The bold values represent the highest PSNR (dB) achieved at specific subrate for video sequence.
Table 5. Average PNSR evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different HD video sequences at various subrates having GoP = 8.
Table 5. Average PNSR evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different HD video sequences at various subrates having GoP = 8.
Book Arrival
Subrate0.050.100.150.200.250.30
BCS-TV-AL327.2430.2132.0833.6434.7735.8
JD-TV32.9933.835.0436.1936.9237.76
Gain5.753.592.962.552.151.96
Newspaper
Subrate0.050.100.150.200.250.30
BCS-TV-AL326.0029.5831.7133.3034.6440.13
JD-TV32.5835.0636.3637.3238.2338.80
Gain6.585.484.654.023.593.33
Lovebird
Subrate0.050.100.150.200.250.30
BCS-TV-AL325.0828.6430.3532.0233.5434.86
JD-TV34.5036.5237.9538.9440.0840.12
Gain9.428.887.606.926.545.26
Note: The bold values represent the PSNR (dB) gain difference achieved by JD-TV for various video sequence.
Table 6. Average SSIM evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different HD at various subrates having GoP = 8.
Table 6. Average SSIM evaluation of proposed JD-TV scheme and the independent BCS-TV-AL3 for different HD at various subrates having GoP = 8.
Book Arrival
Subrate0.050.100.150.200.250.30
BCS-TV-AL30.790.850.880.90.90.92
JD-TV0.910.920.940.950.960.97
Gain0.120.070.060.050.060.05
Newspaper
Subrate0.050.100.150.200.250.30
BCS-TV-AL30.770.840.880.910.920.93
JD-TV0.930.940.950.970.970.98
Gain0.160.100.070.060.050.05
Lovebird
Subrate0.050.100.150.200.250.30
BCS-TV-AL30.750.830.870.90.910.93
JD-TV0.950.960.960.970.970.98
Gain0.200.130.090.070.060.05
Note: The bold values represent the PSNR (dB) gain difference achieved by JD-TV for various video sequence.
Table 7. Average PSNR gain (dB) analysis of the proposed JD with conventional CS schemes for various video sequences.
Table 7. Average PSNR gain (dB) analysis of the proposed JD with conventional CS schemes for various video sequences.
Subrate0.050.100.150.200.250.30
Hall MonitorJD-TV6.886.756.686.516.476.33
MC-BCS-SPL2.643.884.965.775.986.21
kt-Focuss1.501.962.453.213.974.05
MS-Residual0.851.061.552.172.883.35
Mother DaughterJD-TV5.965.835.695.575.415.27
MC-BCS-SPL2.173.043.774.304.855.14
kt-Focuss1.081.872.653.343.994.75
MS-Residual0.380.771.271.832.593.05
Coast GuardJD-TV2.762.652.552.422.312.22
MC-BCS-SPL0.951.351.491.741.982.15
kt-Focuss0.451.011.211.401.601.89
MS-Residual0.250.540.750.951.201.49
ForemanJD-TV5.564.964.263.692.142.08
MC-BCS-SPL0.902.012.813.443.573.74
kt-Focuss0.450.670.811.071.231.59
MS-Residual0.150.240.550.770.951.19
Mobile CalendarJD-TV3.843.73.583.433.33.17
MC-BCS-SPL1.071.932.263.063.924.66
kt-Focuss0.771.1322.633.233.97
MS-Residual0.370.661.061.632.132.89
StefanJD-TV4.284.154.124.073.983.84
MC-BCS-SPL0.350.550.931.462.182.91
kt-Focuss0.230.410.671.11.532.06
MS-Residual0.160.211.060.550.981.15
Note: The bold values represent the highest PSNR (dB) gain achieved at specific subrate for video sequence.
Table 8. Analysis of the proposed JD and BCS-TV-AL3 scheme in terms of bits saved at various reconstruction qualities (PSNR).
Table 8. Analysis of the proposed JD and BCS-TV-AL3 scheme in terms of bits saved at various reconstruction qualities (PSNR).
 
PSNR
Bits
BCS-TV-AL3
Bits
JD-TV
Bits
Saved
Bits
Saving (%)
Bits
Average (%)
Hall Monitor~24.031306819801108884
~25.05174245148122767275
~25.572019671281306868
Coast Guard~23.42122765940633652
~24.0214256910851483740
~24.531742412276514830
Mother Daughter~29.3791083960514858
~30.2013068712859404549
~31.59174249908751643
Calendar~18.3291083168594066
~18.7812276514871285858
~19.39142567128712850
Table 9. Average encoding (execution) time (sec) comparison of proposed CS based JD-TV codec with conventional video codecs (DISCOVER, H.264 (intra)) for various video sequences.
Table 9. Average encoding (execution) time (sec) comparison of proposed CS based JD-TV codec with conventional video codecs (DISCOVER, H.264 (intra)) for various video sequences.
SequenceProposed (JD-TV)DISCOVERH.264 (Intra)
Hall-Monitor6.31 s31.21 s65.66 s
Mother Daughter6.39 s40.65 s86.58 s
Coast-Guard6.53 s31.05 s62.40 s
Foreman6.94 s36.40 s78.50 s
Mobile Calendar 7.38 s60.08 s125.45 s
Table 10. Average decoding (execution) time (sec) comparison of proposed CS based JD-TV codec with conventional video codecs (DISCOVER, H.264 (intra)) for various video sequences.
Table 10. Average decoding (execution) time (sec) comparison of proposed CS based JD-TV codec with conventional video codecs (DISCOVER, H.264 (intra)) for various video sequences.
SequenceProposed (JD-TV)DISCOVERH.264 (Intra)
Hall-Monitor7.21 s1.13 s0.60 s
Mother Daughter7.17 s1.11 s0.59 s
Coast-Guard9.22 s1.27 s0.75 s
Foreman9.19 s1.22 s0.67 s
Mobile Calendar 9.43 s1.27 s0.75 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Ebrahim, M.; Adil, S.H.; Raza, K.; Ali, S.S.A. Block Compressive Sensing Single-View Video Reconstruction Using Joint Decoding Framework for Low Power Real Time Applications. Appl. Sci. 2020, 10, 7963. https://doi.org/10.3390/app10227963

AMA Style

Ebrahim M, Adil SH, Raza K, Ali SSA. Block Compressive Sensing Single-View Video Reconstruction Using Joint Decoding Framework for Low Power Real Time Applications. Applied Sciences. 2020; 10(22):7963. https://doi.org/10.3390/app10227963

Chicago/Turabian Style

Ebrahim, Mansoor, Syed Hasan Adil, Kamran Raza, and Syed Saad Azhar Ali. 2020. "Block Compressive Sensing Single-View Video Reconstruction Using Joint Decoding Framework for Low Power Real Time Applications" Applied Sciences 10, no. 22: 7963. https://doi.org/10.3390/app10227963

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop