[go: up one dir, main page]
More Web Proxy on the site http://driver.im/
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

Article Types

Countries / Regions

Search Results (8)

Search Parameters:
Keywords = single-block encoding scheme

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
12 pages, 381 KiB  
Article
Group Testing with Blocks of Positives and Inhibitors
by Thach V. Bui, Isao Echizen, Minoru Kuribayashi, Tetsuya Kojima and Thuc D. Nguyen
Entropy 2022, 24(11), 1562; https://doi.org/10.3390/e24111562 - 30 Oct 2022
Viewed by 1870
Abstract
The main goal of group testing is to identify a small number of specific items among a large population of items. In this paper, we consider specific items as positives and inhibitors and non-specific items as negatives. In particular, we consider a novel [...] Read more.
The main goal of group testing is to identify a small number of specific items among a large population of items. In this paper, we consider specific items as positives and inhibitors and non-specific items as negatives. In particular, we consider a novel model called group testing with blocks of positives and inhibitors. A test on a subset of items is positive if the subset contains at least one positive and does not contain any inhibitors, and it is negative otherwise. In this model, the input items are linearly ordered, and the positives and inhibitors are subsets of small blocks (at unknown locations) of consecutive items over that order. We also consider two specific instantiations of this model. The first instantiation is that model that contains a single block of consecutive items consisting of exactly known numbers of positives and inhibitors. The second instantiation is the model that contains a single block of consecutive items containing known numbers of positives and inhibitors. Our contribution is to propose efficient encoding and decoding schemes such that the numbers of tests used to identify only positives or both positives and inhibitors are less than the ones in the state-of-the-art schemes. Moreover, the decoding times mostly scale to the numbers of tests that are significantly smaller than the state-of-the-art ones, which scale to both the number of tests and the number of items. Full article
(This article belongs to the Special Issue Theory and Applications of Information Processing Algorithms)
Show Figures

Figure 1

Figure 1
<p>Three models for blocks of positives and inhibitors. Red, purple, and black dots represent positives, inhibitors, and negative items, respectively. A double arrow line stands for a block of <span class="html-italic">D</span> consecutive items. The first, second, and third models are a single block of consecutive positives and inhibitors, single block of positives and inhibitors, and blocks of positives and inhibitors. The second model is a generalization of the first model, and the third model is a generalization of the first two models.</p>
Full article ">
9 pages, 654 KiB  
Article
Implementation of Quantum Algorithms via Fast Three-Rydberg-Atom CCZ Gates
by Shiqing Tang, Chong Yang, Dongxiao Li and Xiaoqiang Shao
Entropy 2022, 24(10), 1371; https://doi.org/10.3390/e24101371 - 27 Sep 2022
Cited by 2 | Viewed by 2270
Abstract
Multiqubit CCZ gates form one of the building blocks of quantum algorithms and have been involved in achieving many theoretical and experimental triumphs. Designing a simple and efficient multiqubit gate for quantum algorithms is still by no means trivial as the number of [...] Read more.
Multiqubit CCZ gates form one of the building blocks of quantum algorithms and have been involved in achieving many theoretical and experimental triumphs. Designing a simple and efficient multiqubit gate for quantum algorithms is still by no means trivial as the number of qubits increases. Here, by virtue of the Rydberg blockade effect, we propose a scheme to rapidly implement a three-Rydberg-atom CCZ gate via a single Rydberg pulse, and successfully apply the gate to realize the three-qubit refined Deutsch–Jozsa algorithm and three-qubit Grover search. The logical states of the three-qubit gate are encoded to the same ground states to avoid an adverse effect of the atomic spontaneous emission. Furthermore, there is no requirement for individual addressing of atoms in our protocol. Full article
(This article belongs to the Special Issue Quantum Computing for Complex Dynamics)
Show Figures

Figure 1

Figure 1
<p>Schematic illustrations for the setup and atomic levels of the three-Rydberg-atom system. Each atom includes two ground states <math display="inline"><semantics> <mrow> <mo>|</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>〉</mo> </mrow> </semantics></math> and one Rydberg state <math display="inline"><semantics> <mrow> <mo>|</mo> <mi>r</mi> <mo>〉</mo> </mrow> </semantics></math>. The Rydberg state is dispersively coupled to the ground states via one common Rydberg pulse with effective Rabi frequency <math display="inline"><semantics> <mrow> <mi mathvariant="normal">Ω</mi> <mo>≈</mo> <mrow> <mn>2</mn> <mi>π</mi> <mo>×</mo> <mn>3.5</mn> </mrow> </mrow> </semantics></math> MHz and adjustable detuning <math display="inline"><semantics> <mi>δ</mi> </semantics></math>. The Rydberg–Rydberg interaction between the <span class="html-italic">i</span>- and <span class="html-italic">j</span>-th atoms is described as <math display="inline"><semantics> <msub> <mi>U</mi> <mrow> <mi>i</mi> <mi>j</mi> </mrow> </msub> </semantics></math>.</p>
Full article ">Figure 2
<p>(<b>a</b>) The average gate fidelity with respect to <math display="inline"><semantics> <mi>δ</mi> </semantics></math> as well as the gate operation time, where the system is governed by the full Hamiltonian of Equation (<a href="#FD1-entropy-24-01371" class="html-disp-formula">1</a>) to realize the gate of Equation (<a href="#FD4-entropy-24-01371" class="html-disp-formula">4</a>). (<b>b</b>) The average gate fidelity without the three single qubit logical gates as functions of <math display="inline"><semantics> <mi>δ</mi> </semantics></math> and <span class="html-italic">t</span>. For the two sub-pictures, the Rabi frequencies and the interaction strengths are <math display="inline"><semantics> <mrow> <mi mathvariant="normal">Ω</mi> <mo>=</mo> <mn>2</mn> <mi>π</mi> <mo>×</mo> <mn>3.5</mn> </mrow> </semantics></math> MHz and <math display="inline"><semantics> <mrow> <msub> <mi>U</mi> <mrow> <mi>i</mi> <mi>j</mi> </mrow> </msub> <mo>=</mo> <mi>U</mi> <mo>≈</mo> <mn>2</mn> <mi>π</mi> <mo>×</mo> <mn>35</mn> </mrow> </semantics></math> MHz corresponding the atomic distance <math display="inline"><semantics> <mrow> <msub> <mi>r</mi> <mrow> <mi>i</mi> <mi>j</mi> </mrow> </msub> <mo>=</mo> <mi>r</mi> <mo>≈</mo> <mn>5.4</mn> </mrow> </semantics></math> <math display="inline"><semantics> <mi mathvariant="sans-serif">μ</mi> </semantics></math>m.</p>
Full article ">Figure 3
<p>Contour plot of average gate fidelity of <math display="inline"><semantics> <msubsup> <mi>U</mi> <mrow> <mi>B</mi> <mn>1</mn> </mrow> <mn>3</mn> </msubsup> </semantics></math> with respect to the detuning and the gate operation time. The relevant parameters are the same as those of <a href="#entropy-24-01371-f002" class="html-fig">Figure 2</a>a.</p>
Full article ">Figure 4
<p>Fidelity of the state searched for as functions of the iteration number with different Rydberg–Rydberg interaction strength. The marked state is <math display="inline"><semantics> <mrow> <mo>|</mo> <mn>101</mn> <mo>〉</mo> </mrow> </semantics></math> and the initial state is <math display="inline"><semantics> <mrow> <mo>|</mo> <mn>000</mn> <mo>〉</mo> </mrow> </semantics></math>. The relevant parameters are the same as those of <a href="#entropy-24-01371-f002" class="html-fig">Figure 2</a>a and the gate operation time for <math display="inline"><semantics> <mrow> <mi mathvariant="script">C</mi> <mo>(</mo> <mi>t</mi> <mo>)</mo> </mrow> </semantics></math> is set as <math display="inline"><semantics> <mrow> <mn>0.8049</mn> </mrow> </semantics></math> <math display="inline"><semantics> <mi mathvariant="sans-serif">μ</mi> </semantics></math>s. For simplicity, the Rydberg–Rydberg interaction strength between the <span class="html-italic">i</span>- and <span class="html-italic">j</span>-th atom are assumed as <math display="inline"><semantics> <mrow> <msub> <mi>U</mi> <mrow> <mi>i</mi> <mi>j</mi> </mrow> </msub> <mo>=</mo> <mi>U</mi> </mrow> </semantics></math>.</p>
Full article ">
27 pages, 1358 KiB  
Article
Bagged Tree and ResNet-Based Joint End-to-End Fast CTU Partition Decision Algorithm for Video Intra Coding
by Yixiao Li, Lixiang Li, Yuan Fang, Haipeng Peng and Nam Ling
Electronics 2022, 11(8), 1264; https://doi.org/10.3390/electronics11081264 - 16 Apr 2022
Cited by 10 | Viewed by 2437
Abstract
Video coding standards, such as high-efficiency video coding (HEVC), versatile video coding (VVC), and AOMedia video 2 (AV2), achieve an optimal encoding performance by traversing all possible combinations of coding unit (CU) partition and selecting the combination with the minimum coding cost. It [...] Read more.
Video coding standards, such as high-efficiency video coding (HEVC), versatile video coding (VVC), and AOMedia video 2 (AV2), achieve an optimal encoding performance by traversing all possible combinations of coding unit (CU) partition and selecting the combination with the minimum coding cost. It is still necessary to further reduce the encoding time of HEVC, because HEVC is one of the most widely used coding standards. In HEVC, the process of searching for the best performance is the source of most of the encoding complexity. To reduce the complexity of the coding block partition in HEVC, a new end-to-end fast algorithm is presented to aid the partition structure decisions of the coding tree unit (CTU) in intra coding. In the proposed method, the partition structure decision problem of a CTU is solved by a novel two-stage strategy. In the first stage, a bagged tree model is employed to predict the splitting of a CTU. In the second stage, the partition problem of a 32 × 32-sized CU is modeled as a 17-output classification task for the first time, so that it can be solved by a single prediction. To achieve a high prediction accuracy, a residual network (ResNet) with 34 layers is employed. Jointly using bagged tree and ResNet, the proposed fast CTU partition algorithm is able to generate the partition quad-tree structure of a CTU through an end-to-end prediction process, which abandons the traditional scheme of making multiple decisions at various depth levels. In addition, several datasets are used in this paper to lay the foundation for high prediction accuracy. Compared with the original HM16.7 encoder, the experimental results show that the proposed algorithm can reduce the encoding time by 60.29% on average, while the Bjøntegaard delta rate (BD-rate) loss is as low as 2.03%, which outperforms the results of most of the state-of-the-art approaches in the field of fast intra CU partition. Full article
Show Figures

Figure 1

Figure 1
<p>Recursive partition process of a CTU in the HEVC intra coding.</p>
Full article ">Figure 2
<p>A partition example of a CTU and the corresponding quad-tree splitting structure in the HEVC intra coding.</p>
Full article ">Figure 3
<p>The structure of a bagged tree model. An illustration of how the final prediction for a sample input is generated by such a model, which is composed of multiple decision trees, each trained on a random subset of data.</p>
Full article ">Figure 4
<p>Structure of a building block in ResNet and two examples of a basic block for ResNet constructions of different depths [<a href="#B49-electronics-11-01264" class="html-bibr">49</a>]. (<b>a</b>) The basic structure with a shortcut for ResNet. (<b>b</b>) An example of a basic block used to build a shallow ResNet. (<b>c</b>) The “bottleneck” used to construct a deep ResNet by stacking.</p>
Full article ">Figure 5
<p>Flowchart of the proposed bagged tree and ResNet-based joint fast CTU partition algorithm.</p>
Full article ">Figure 6
<p>An example of CTU partition structure determination process with the corresponding partition quad-tree. <math display="inline"><semantics> <mrow> <mi>P</mi> <mi>L</mi> </mrow> </semantics></math> is the predicted label for a 32 × 32 CU output by ResNet. <math display="inline"><semantics> <mrow> <mi>P</mi> <mi>L</mi> </mrow> </semantics></math> is set as an integer from 1 to 17.</p>
Full article ">Figure 7
<p>Sobel filters for different directions. (<b>a</b>) Example of a 3 × 3 square pixel block to be filtered by Sobel filters. (<b>b</b>) Sobel filter for horizontal direction. (<b>c</b>) Sobel filter for vertical direction. (<b>d</b>,<b>e</b>) Sobel filters for 45<math display="inline"><semantics> <msup> <mrow/> <mo>∘</mo> </msup> </semantics></math> and 135<math display="inline"><semantics> <msup> <mrow/> <mo>∘</mo> </msup> </semantics></math>, respectively.</p>
Full article ">Figure 8
<p>Haar filters for different directions. (<b>a</b>) Example of a 2 × 2 square pixel block to be filtered by Haar filters. (<b>b</b>) Haar filter for horizontal direction. (<b>c</b>) Haar filter for vertical direction. (<b>d</b>) Haar filter for diagonal direction.</p>
Full article ">Figure 9
<p>Three filters used for interest point detection in horizontal, vertical, and diagonal directions. (<b>a</b>) The filter used in the horizontal direction. (<b>b</b>) The filter used in the vertical direction. (<b>c</b>) The filter used in the diagonal direction. (<b>a</b>–<b>c</b>) are performed on each pixel of a CTU, and the corresponding filtering responses are <math display="inline"><semantics> <msub> <mi>D</mi> <mrow> <mi>x</mi> <mi>x</mi> </mrow> </msub> </semantics></math>, <math display="inline"><semantics> <msub> <mi>D</mi> <mrow> <mi>y</mi> <mi>y</mi> </mrow> </msub> </semantics></math>, and <math display="inline"><semantics> <msub> <mi>D</mi> <mrow> <mi>x</mi> <mi>y</mi> </mrow> </msub> </semantics></math>, respectively.</p>
Full article ">Figure 10
<p>Architecture of the ResNet used in this paper. The dotted lines represent the shortcuts which increase the dimensions.</p>
Full article ">Figure 11
<p>Class labels and corresponding partition structures of a 32 × 32 CU for ResNet multi-classification. Variate <math display="inline"><semantics> <mrow> <mi>l</mi> <mi>a</mi> <mi>b</mi> <mi>e</mi> <mi>l</mi> </mrow> </semantics></math> is the class index of different partitions for a CU of size 32 × 32.</p>
Full article ">Figure 12
<p>Partition results of the 180th frame in the sequence <span class="html-italic">RaceHorses</span> (416 × 240), which is encoded by the original HM 16.7 with QP 22. The black line represents the splitting boundaries between CUs.</p>
Full article ">Figure 13
<p>Partition results of the 180th frame in the sequence <span class="html-italic">RaceHorses</span> (416 × 240), which is encoded by the third version <math display="inline"><semantics> <msub> <mi>BTRNFA</mi> <mi>joint</mi> </msub> </semantics></math> of the proposed algorithm with QP 22. The black line represents the same partition results as those of the original HM16.7. The gold line represents the boundaries of CUs, which are split by the original HM16.7 but are not split by the <math display="inline"><semantics> <msub> <mi>BTRNFA</mi> <mi>joint</mi> </msub> </semantics></math>. The boundaries of CUs, which are non-split by the original HM16.7 but are split by the <math display="inline"><semantics> <msub> <mi>BTRNFA</mi> <mi>joint</mi> </msub> </semantics></math>, are shown with a red line.</p>
Full article ">
15 pages, 4180 KiB  
Technical Note
Integrating EfficientNet into an HAFNet Structure for Building Mapping in High-Resolution Optical Earth Observation Data
by Luca Ferrari, Fabio Dell’Acqua, Peng Zhang and Peijun Du
Remote Sens. 2021, 13(21), 4361; https://doi.org/10.3390/rs13214361 - 29 Oct 2021
Cited by 11 | Viewed by 2678
Abstract
Automated extraction of buildings from Earth observation (EO) data is important for various applications, including updating of maps, risk assessment, urban planning, and policy-making. Combining data from different sensors, such as high-resolution multispectral images (HRI) and light detection and ranging (LiDAR) data, has [...] Read more.
Automated extraction of buildings from Earth observation (EO) data is important for various applications, including updating of maps, risk assessment, urban planning, and policy-making. Combining data from different sensors, such as high-resolution multispectral images (HRI) and light detection and ranging (LiDAR) data, has shown great potential in building extraction. Deep learning (DL) is increasingly used in multi-modal data fusion and urban object extraction. However, DL-based multi-modal fusion networks may under-perform due to insufficient learning of “joint features” from multiple sources and oversimplified approaches to fusing multi-modal features. Recently, a hybrid attention-aware fusion network (HAFNet) has been proposed for building extraction from a dataset, including co-located Very-High-Resolution (VHR) optical images and light detection and ranging (LiDAR) joint data. The system reported good performances thanks to the adaptivity of the attention mechanism to the features of the information content of the three streams but suffered from model over-parametrization, which inevitably leads to long training times and heavy computational load. In this paper, the authors propose a restructuring of the scheme, which involved replacing VGG-16-like encoders with the recently proposed EfficientNet, whose advantages counteract exactly the issues found with the HAFNet scheme. The novel configuration was tested on multiple benchmark datasets, reporting great improvements in terms of processing times, and also in terms of accuracy. The new scheme, called HAFNetE (HAFNet with EfficientNet integration), appears indeed capable of achieving good results with less parameters, translating into better computational efficiency. Based on these findings, we can conclude that, given the current advancements in single-thread schemes, the classical multi-thread HAFNet scheme could be effectively transformed by the HAFNetE scheme by replacing VGG-16 with EfficientNet blocks on each single thread. The remarkable reduction achieved in computational requirements moves the system one step closer to on-board implementation in a possible, future “urban mapping” satellite constellation. Full article
(This article belongs to the Special Issue Deep Learning for Remote Sensing Data)
Show Figures

Figure 1

Figure 1
<p>Harmful information in input data. (<b>a</b>) RGB patch containing discriminative information. (<b>b</b>) DSM patch containing incorrect information. (<b>c</b>) Ground truth map. (<b>d</b>) Segmentation result. The Att-MFBlock re-weights the RGB and the DSM input so that RGB information is highlighted and the damaged DSM information is suppressed.</p>
Full article ">Figure 2
<p>Scheme of the HAFNetE network.</p>
Full article ">Figure 3
<p>(<b>a</b>) RGB image patch. (<b>b</b>) DSM patch. (<b>c</b>) Corresponding binary thematic map. Building pixels are displayed in white, whereas non-building pixels are displayed in black.</p>
Full article ">Figure 4
<p>Example of visible distortion in RGB input images.</p>
Full article ">Figure 5
<p>Training (blue) and validation (orange) curves obtained using the Adam optimizer. From top-left, clockwise, the four graphs represent the measures of loss, fscore, IoU, and accuracy, respectively.</p>
Full article ">Figure 6
<p>(<b>a</b>) Input RGB patches. (<b>b</b>) Input DSM patches. (<b>c</b>) Model soft predictions. (<b>d</b>) Thresholded predictions. (<b>e</b>) Label patches. Please note that the corrupted DSM input (<b>b</b>) is adaptively re-weighted by the Att-MFBlock, thus suppressing misleading information. Thanks to this mechanism, a final correct segmentation result is produced.</p>
Full article ">Figure 7
<p>Two examples of results obtained on two entire 6000 × 6000 pixel tiles from the Potsdam dataset. (<b>a</b>) RGB TOP image of Tile 1. (<b>b</b>) HAFNetE classification result on Tile 1. (<b>c</b>) Ground Truth map for Tile 1. (<b>d</b>) RGB TOP image of Tile 2. (<b>e</b>) HAFNetE classification result on Tile 2. (<b>f</b>) Ground Truth map for Tile 2.</p>
Full article ">
24 pages, 16451 KiB  
Article
A Flexible Coding Scheme Based on Block Krylov Subspace Approximation for Light Field Displays with Stacked Multiplicative Layers
by Joshitha Ravishankar, Mansi Sharma and Pradeep Gopalakrishnan
Sensors 2021, 21(13), 4574; https://doi.org/10.3390/s21134574 - 4 Jul 2021
Cited by 12 | Viewed by 3937
Abstract
To create a realistic 3D perception on glasses-free displays, it is critical to support continuous motion parallax, greater depths of field, and wider fields of view. A new type of Layered or Tensor light field 3D display has attracted greater attention these days. [...] Read more.
To create a realistic 3D perception on glasses-free displays, it is critical to support continuous motion parallax, greater depths of field, and wider fields of view. A new type of Layered or Tensor light field 3D display has attracted greater attention these days. Using only a few light-attenuating pixelized layers (e.g., LCD panels), it supports many views from different viewing directions that can be displayed simultaneously with a high resolution. This paper presents a novel flexible scheme for efficient layer-based representation and lossy compression of light fields on layered displays. The proposed scheme learns stacked multiplicative layers optimized using a convolutional neural network (CNN). The intrinsic redundancy in light field data is efficiently removed by analyzing the hidden low-rank structure of multiplicative layers on a Krylov subspace. Factorization derived from Block Krylov singular value decomposition (BK-SVD) exploits the spatial correlation in layer patterns for multiplicative layers with varying low ranks. Further, encoding with HEVC eliminates inter-frame and intra-frame redundancies in the low-rank approximated representation of layers and improves the compression efficiency. The scheme is flexible to realize multiple bitrates at the decoder by adjusting the ranks of BK-SVD representation and HEVC quantization. Thus, it would complement the generality and flexibility of a data-driven CNN-based method for coding with multiple bitrates within a single training framework for practical display applications. Extensive experiments demonstrate that the proposed coding scheme achieves substantial bitrate savings compared with pseudo-sequence-based light field compression approaches and state-of-the-art JPEG and HEVC coders. Full article
(This article belongs to the Collection Machine Learning for Multimedia Communications)
Show Figures

Figure 1

Figure 1
<p>Light field is defined in 4-D space. The structure (<b>a</b>) and configuration (<b>b</b>) of layered light field display for constructing multiplicative layers are shown in 2-D for simplicity; (<b>c</b>) the light ray is parameterized by point of intersection with the <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>u</mi> <mo>,</mo> <mi>v</mi> <mo>)</mo> </mrow> </semantics></math> plane and the <math display="inline"><semantics> <mrow> <mo>(</mo> <mi>s</mi> <mo>,</mo> <mi>t</mi> <mo>)</mo> </mrow> </semantics></math> plane located at a depth <span class="html-italic">z</span>.</p>
Full article ">Figure 2
<p>The complete workflow of the proposed scheme comprising of three prime components: conversion of light field views into multiplicative layers, low-rank approximation of layers and HEVC encoding, and the decoding of the approximated layers followed by the reconstruction of the light field.</p>
Full article ">Figure 3
<p>The three optimal multiplicative layers obtained from CNN (with 20 convolutional layers, trained for 20 epochs with a learning rate of 0.0001, and batch size 15) for <span class="html-italic">Bunnies</span> data (<span class="html-italic">BLOCK I</span> of the proposed scheme).</p>
Full article ">Figure 4
<p>The BK-SVD procedure adopted in the proposed scheme. The layers <math display="inline"><semantics> <msub> <mi>M</mi> <mi>z</mi> </msub> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>z</mi> <mo>∈</mo> <mo>{</mo> <mo>−</mo> <mn>1</mn> <mo>,</mo> <mn>0</mn> <mo>,</mo> <mn>1</mn> <mo>}</mo> </mrow> </semantics></math> are split into color channels and rearranged to form matrices <math display="inline"><semantics> <msup> <mi>B</mi> <mrow> <mi>c</mi> <mi>h</mi> </mrow> </msup> </semantics></math>, <math display="inline"><semantics> <mrow> <mi>c</mi> <mi>h</mi> <mo>∈</mo> <mo>{</mo> <mi>r</mi> <mo>,</mo> <mi>g</mi> <mo>,</mo> <mi>b</mi> <mo>}</mo> </mrow> </semantics></math>. The rank <span class="html-italic">k</span> approximation of <math display="inline"><semantics> <msup> <mi>B</mi> <mrow> <mi>c</mi> <mi>h</mi> </mrow> </msup> </semantics></math> results in <math display="inline"><semantics> <msup> <mi>W</mi> <mrow> <mi>c</mi> <mi>h</mi> </mrow> </msup> </semantics></math>, which are then divided and rearranged to attain the approximated layers.</p>
Full article ">Figure 5
<p>The workflow of the encoding and decoding steps of HEVC codec.</p>
Full article ">Figure 6
<p>For the central viewpoint <math display="inline"><semantics> <mrow> <mrow> <mo>(</mo> <msup> <mi>s</mi> <mo>*</mo> </msup> <mo>,</mo> <msup> <mi>t</mi> <mo>*</mo> </msup> <mo>)</mo> </mrow> <mo>=</mo> <mrow> <mo>(</mo> <mn>0</mn> <mo>,</mo> <mn>0</mn> <mo>)</mo> </mrow> </mrow> </semantics></math>, the decoded layers <math display="inline"><semantics> <msub> <mover accent="true"> <mi>M</mi> <mo>‵</mo> </mover> <mi>z</mi> </msub> </semantics></math> are translated to <math display="inline"><semantics> <msub> <mover> <mi>M</mi> <mo>‶</mo> </mover> <mrow> <mi>z</mi> <mo>(</mo> <mn>0</mn> <mo>,</mo> <mn>0</mn> <mo>)</mo> </mrow> </msub> </semantics></math>. The corresponding color channels are multiplied element-wise to obtain <math display="inline"><semantics> <msubsup> <mi>I</mi> <mrow> <mo>(</mo> <mn>0</mn> <mo>,</mo> <mn>0</mn> <mo>)</mo> </mrow> <mrow> <mi>c</mi> <mi>h</mi> </mrow> </msubsup> </semantics></math>. The final central light field view <math display="inline"><semantics> <msub> <mi>I</mi> <mrow> <mo>(</mo> <mn>0</mn> <mo>,</mo> <mn>0</mn> <mo>)</mo> </mrow> </msub> </semantics></math> is obtained by combining the red, blue, and green color channels.</p>
Full article ">Figure 7
<p>(<b>a</b>) The extracted 13 × 13 views of the <span class="html-italic">Fountain-Vincent2</span>; central view of (<b>b</b>) <span class="html-italic">Bikes</span>; (<b>c</b>) <span class="html-italic">Fountain-Vincent2</span>; (<b>d</b>) <span class="html-italic">Stone-Pillars Outside</span>.</p>
Full article ">Figure 8
<p>(<b>a</b>) Training results of the CNN with 20 convolutional layers. Optimal PSNR is observed for model run for 20 epochs with learning rate 0.0001 and batch size 15; (<b>b</b>–<b>d</b>) rate-distortion curves comparing results of CNN structures with 15, 20, and 25 convolutional layers trained for 20 epochs with learning rate 0.0001 and batch size 15.</p>
Full article ">Figure 9
<p>The three multiplicative layers of <span class="html-italic">Bikes</span>, <span class="html-italic">Fountain-Vincent2</span>, and <span class="html-italic">Stone-Pillars Outside</span> light fields obtained from the convolutional neural network.</p>
Full article ">Figure 10
<p>Comparison of the original and reconstructed views. (<b>a</b>) Original central views of the three datasets; (<b>b</b>) reconstructed central views using our proposed scheme for rank 20 and QP 2; (<b>c</b>) reconstructed central views using our proposed scheme for rank 60 and QP 2.</p>
Full article ">Figure 11
<p>(<b>a</b>–<b>c</b>) The bitrate versus PSNR graphs of Ahmad et al. coding scheme for all three datasets; (<b>d</b>) graph depicting the maximal bytes per allocations performed against the PSNR for <span class="html-italic">Bikes</span>, <span class="html-italic">Fountain-Vincent2</span>, and <span class="html-italic">Stone-Pillars Outside</span> datasets using the JPEG codec.</p>
Full article ">Figure 12
<p>Rate-distortion curves for the proposed compression scheme and HEVC codec for the three datasets.</p>
Full article ">Figure 13
<p>Comparative mean SSIM for the proposed scheme and anchor coding methods. Average SSIM was computed over all 169 views and all quantization parameters.</p>
Full article ">Figure 14
<p>Mean SSIM scores over each QP of decoded views in <span class="html-italic">BLOCK III</span> of proposed scheme with <span class="html-italic">(ML(3))</span> and without <span class="html-italic">(AV(169))</span> multiplicative layers. Experiment evaluated on <span class="html-italic">Bikes</span> light field for BK-SVD ranks 20 and 60.</p>
Full article ">Figure 15
<p>Computation time and accuracy of reproduced light fields using analytical and CNN-based optimization of multiplicative layers.</p>
Full article ">Figure 16
<p>View 19 of <span class="html-italic">Bunnies</span> reproduced using analytical (ANA) and CNN-based (CNN) optimization of multiplicative layers, with corresponding difference images. (<b>a</b>) ANA: Reproduced view and error, PSNR: 19.94 dB, SSIM: 0.895; (<b>b</b>) CNN: Reproduced view and error, PSNR: 22.18 dB, SSIM: 0.918.</p>
Full article ">
23 pages, 5486 KiB  
Article
Block Compressive Sensing Single-View Video Reconstruction Using Joint Decoding Framework for Low Power Real Time Applications
by Mansoor Ebrahim, Syed Hasan Adil, Kamran Raza and Syed Saad Azhar Ali
Appl. Sci. 2020, 10(22), 7963; https://doi.org/10.3390/app10227963 - 10 Nov 2020
Cited by 2 | Viewed by 1802
Abstract
Several real-time visual monitoring applications such as surveillance, mental state monitoring, driver drowsiness and patient care, require equipping high-quality cameras with wireless sensors to form visual sensors and this creates an enormous amount of data that has to be managed and transmitted at [...] Read more.
Several real-time visual monitoring applications such as surveillance, mental state monitoring, driver drowsiness and patient care, require equipping high-quality cameras with wireless sensors to form visual sensors and this creates an enormous amount of data that has to be managed and transmitted at the sensor node. Moreover, as the sensor nodes are battery-operated, power utilization is one of the key concerns that must be considered. One solution to this issue is to reduce the amount of data that has to be transmitted using specific compression techniques. The conventional compression standards are based on complex encoders (which require high processing power) and simple decoders and thus are not pertinent for battery-operated applications, i.e., VSN (primitive hardware). In contrast, compressive sensing (CS) a distributive source coding mechanism, has transformed the standard coding mechanism and is based on the idea of a simple encoder (i.e., transmitting fewer data-low processing requirements) and a complex decoder and is considered a better option for VSN applications. In this paper, a CS-based joint decoding (JD) framework using frame prediction (using keyframes) and residual reconstruction for single-view video is proposed. The idea is to exploit the redundancies present in the key and non-key frames to produce side information to refine the non-key frames’ quality. The proposed method consists of two main steps: frame prediction and residual reconstruction. The final reconstruction is performed by adding a residual frame with the predicted frame. The proposed scheme was validated on various arrangements. The association among correlated frames and compression performance is also analyzed. Various arrangements of the frames have been studied to select the one that produces better results. The comprehensive experimental analysis proves that the proposed JD method performs notably better than the independent block compressive sensing scheme at different subrates for various video sequences with low, moderate and high motion contents. Also, the proposed scheme outperforms the conventional CS video reconstruction schemes at lower subrates. Further, the proposed scheme was quantized and compared with conventional video codecs (DISCOVER, H-263, H264) at various bitrates to evaluate its efficiency (rate-distortion, encoding, decoding). Full article
(This article belongs to the Special Issue Advances in Signal, Image and Video Processing)
Show Figures

Figure 1

Figure 1
<p>Archetype of the complete system for single-view video coding using the proposed JD (joint decoding).</p>
Full article ">Figure 2
<p>Correlation analysis of CS measurements among the adjacent frames of various CIF video sequences.</p>
Full article ">Figure 3
<p>Reconstruction of frames using the proposed joint decoding (JD) framework.</p>
Full article ">Figure 4
<p>Visual quality analysis for a complete GoP (J = 8) of News video sequence at a subrate of 0.1 using proposed JD-TV framework.</p>
Full article ">Figure 5
<p>Average SSIM evaluation at GoP sizes of 3, 5, and 8 for different video sequences at different subrates for independent BCS-TV-AL3and proposed JD.</p>
Full article ">Figure 6
<p>Visual quality analysis of different video sequences at multiple subrates using independent BCS-TV-AL3 and the proposed JD-TV.</p>
Full article ">Figure 6 Cont.
<p>Visual quality analysis of different video sequences at multiple subrates using independent BCS-TV-AL3 and the proposed JD-TV.</p>
Full article ">Figure 7
<p>PSNR comparison of proposed SQ-ADPCM-JD with conventional codecs i.e., DISCOVER, H.264 (intra, (I-P-P)) and H.263 (intra, (I-P-P)) at various bitrate for different video sequences.</p>
Full article ">Figure 8
<p>Average execution time (sec) comparison of proposed CS based JD-TV codec with other CS codecs for various video sequences.</p>
Full article ">
14 pages, 2499 KiB  
Article
Formation of Unique Characteristics of Hiding and Encoding of Data Blocks Based on the Fragmented Identifier of Information Processed by Cellular Automata
by Elena Kuleshova, Anatoly Marukhlenko, Vyacheslav Dobritsa and Maxim Tanygin
Computers 2020, 9(2), 51; https://doi.org/10.3390/computers9020051 - 19 Jun 2020
Cited by 5 | Viewed by 3459
Abstract
Currently, the following applications of the theory of cellular automata are known: symmetric encryption, data compression, digital image processing and some others. There are also studies suggesting the possibility of building a public key system based on cellular automata, but this problem has [...] Read more.
Currently, the following applications of the theory of cellular automata are known: symmetric encryption, data compression, digital image processing and some others. There are also studies suggesting the possibility of building a public key system based on cellular automata, but this problem has not been solved. The purpose of the study is to develop an algorithm for hiding and encoding data blocks based on a fragmented identifier of information processed on the basis of cellular automata in the scale of binary data streams using an original method containing an public parameter in the conversion key. A mathematical model of the formation of unique data characteristics is considered, based on the use of patterns that determine the individual neighborhood of elements in cell encryption. A multi-threaded computing scheme has been developed for processing confidential data using the single-key method with a public parameter based on cellular automata and using data segmentation. To study individual chains in data blocks, a software module has been developed that allows one to evaluate the uniformity of information distribution during encryption. A variant of estimating the distribution of bits is proposed that indirectly reflects the cryptographic strength of the method. Based on the developed theoretical principles, a software module is synthesized that implements a transformation rule that takes into account the individual neighborhood of the processed element on the basis of a cellular automata. Experimental studies have shown that this modification made it possible to increase the speed of the method by up to 13 percent due to segmentation and the possibility of parallel processing of the original matrix, as well as to increase cryptographic strength due to the use of a unique chain of pseudo-random neighborhood (hereinafter referred to as PRN) defined by the transformation key. At the same time, it was possible to maintain uniformity of distribution of the output chain at the bit level and to ensure that the number of inversions was included in the confidence interval. Full article
(This article belongs to the Special Issue Selected Papers from MICSECS 2019)
Show Figures

Figure 1

Figure 1
<p>Multi-threaded computing organization scheme.</p>
Full article ">Figure 2
<p>Scheme of the processing system.</p>
Full article ">Figure 3
<p>Relative distribution of data bits, processing using the PRN.</p>
Full article ">Figure 4
<p>Relative distribution of data bits, bit shift and interference in communication lines.</p>
Full article ">Figure 5
<p>Superposition of matrices before and after processing based on (<b>a</b>) dynamic and (<b>b</b>) static neighborhoods.</p>
Full article ">Figure 6
<p>Fragment of a superposition of matrices in the form of a surface.</p>
Full article ">Figure 7
<p>Relative delays associated with the conversion method.</p>
Full article ">Figure 8
<p>Inverse statistics.</p>
Full article ">Figure 9
<p>Inverted bit deviation values (delta).</p>
Full article ">Figure 10
<p>Performance assessment in parallel processing of segments.</p>
Full article ">
18 pages, 1969 KiB  
Article
A New Lossless DNA Compression Algorithm Based on A Single-Block Encoding Scheme
by Deloula Mansouri, Xiaohui Yuan and Abdeldjalil Saidani
Algorithms 2020, 13(4), 99; https://doi.org/10.3390/a13040099 - 20 Apr 2020
Cited by 13 | Viewed by 6211
Abstract
With the emergent evolution in DNA sequencing technology, a massive amount of genomic data is produced every day, mainly DNA sequences, craving for more storage and bandwidth. Unfortunately, managing, analyzing and specifically storing these large amounts of data become a major scientific challenge [...] Read more.
With the emergent evolution in DNA sequencing technology, a massive amount of genomic data is produced every day, mainly DNA sequences, craving for more storage and bandwidth. Unfortunately, managing, analyzing and specifically storing these large amounts of data become a major scientific challenge for bioinformatics. Therefore, to overcome these challenges, compression has become necessary. In this paper, we describe a new reference-free DNA compressor abbreviated as DNAC-SBE. DNAC-SBE is a lossless hybrid compressor that consists of three phases. First, starting from the largest base (Bi), the positions of each Bi are replaced with ones and the positions of other bases that have smaller frequencies than Bi are replaced with zeros. Second, to encode the generated streams, we propose a new single-block encoding scheme (SEB) based on the exploitation of the position of neighboring bits within the block using two different techniques. Finally, the proposed algorithm dynamically assigns the shorter length code to each block. Results show that DNAC-SBE outperforms state-of-the-art compressors and proves its efficiency in terms of special conditions imposed on compressed data, storage space and data transfer rate regardless of the file format or the size of the data. Full article
(This article belongs to the Special Issue Data Compression Algorithms and their Applications)
Show Figures

Figure 1

Figure 1
<p>An overview of the proposed algorithm.</p>
Full article ">Figure 2
<p>The encoding process of the single-block encoding algorithm.</p>
Full article ">Figure 3
<p>The encoding block process using the single-block encoding scheme (SBE).</p>
Full article ">Figure 4
<p>The encoding process of the binary encoding scheme.</p>
Full article ">Figure 5
<p>Workflow of the encoding process of the first block using the SEB scheme.</p>
Full article ">Figure 6
<p>The decoding process of the first compressed code block.</p>
Full article ">
Back to TopTop