[go: up one dir, main page]
More Web Proxy on the site http://driver.im/

CN110557646B - Intelligent inter-view coding method - Google Patents

Intelligent inter-view coding method Download PDF

Info

Publication number
CN110557646B
CN110557646B CN201910775875.1A CN201910775875A CN110557646B CN 110557646 B CN110557646 B CN 110557646B CN 201910775875 A CN201910775875 A CN 201910775875A CN 110557646 B CN110557646 B CN 110557646B
Authority
CN
China
Prior art keywords
prediction
block
mode
prediction block
coding
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Active
Application number
CN201910775875.1A
Other languages
Chinese (zh)
Other versions
CN110557646A (en
Inventor
雷建军
张宗千
侯春萍
张凯明
刘晓寰
石雅南
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Tianjin University
Original Assignee
Tianjin University
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Tianjin University filed Critical Tianjin University
Priority to CN201910775875.1A priority Critical patent/CN110557646B/en
Publication of CN110557646A publication Critical patent/CN110557646A/en
Application granted granted Critical
Publication of CN110557646B publication Critical patent/CN110557646B/en
Active legal-status Critical Current
Anticipated expiration legal-status Critical

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/128Adjusting depth or disparity
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N13/00Stereoscopic video systems; Multi-view video systems; Details thereof
    • H04N13/10Processing, recording or transmission of stereoscopic or multi-view image signals
    • H04N13/106Processing image signals
    • H04N13/161Encoding, multiplexing or demultiplexing different image signal components
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04NPICTORIAL COMMUNICATION, e.g. TELEVISION
    • H04N19/00Methods or arrangements for coding, decoding, compressing or decompressing digital video signals
    • H04N19/50Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding
    • H04N19/597Methods or arrangements for coding, decoding, compressing or decompressing digital video signals using predictive coding specially adapted for multi-view video sequence encoding

Landscapes

  • Engineering & Computer Science (AREA)
  • Multimedia (AREA)
  • Signal Processing (AREA)
  • Compression Or Coding Systems Of Tv Signals (AREA)

Abstract

本发明公开了一种智能视点间的编码方法,包括:分别获取基于相邻块视差矢量模式得到的第一预测块、基于传统视差补偿预测模式得到的第二预测块;构建可变卷积核大小的残差学习卷积神经网络结构;将上述两个预测块的亮度分量通道级联后的结果作为卷积神经网络的输入,网络的输出为原始图像块的增强预测块;根据四种不同的量化参数训练四种不同的网络模型;分别提取第一、第二预测块的亮度分量并进行通道级联,将级联后的结果根据当前编码块的大小以及量化参数值调用相应的网络模型;最后对增强预测块计算率失真代价并与当前编码块的最优模式代价进行比较,如果小于最优模式代价,选择本方法作为最优编码模式,并对相应的标志进行编码传输至解码端。

Figure 201910775875

The invention discloses a coding method between intelligent viewpoints, comprising: respectively acquiring a first prediction block obtained based on a disparity vector mode of adjacent blocks and a second prediction block obtained based on a traditional disparity compensation prediction mode; constructing a variable convolution kernel The convolutional neural network structure is learned from the residual of the size; the concatenated result of the luminance component channels of the above two prediction blocks is used as the input of the convolutional neural network, and the output of the network is the enhanced prediction block of the original image block; according to four different The quantization parameters are used to train four different network models; the luminance components of the first and second prediction blocks are extracted respectively and channel concatenation is performed, and the concatenated result is called according to the size of the current coding block and the quantization parameter value. The corresponding network model is called ; Finally, calculate the rate-distortion cost of the enhanced prediction block and compare it with the optimal mode cost of the current encoding block. If it is less than the optimal mode cost, select this method as the optimal encoding mode, and encode the corresponding flag and transmit it to the decoding end .

Figure 201910775875

Description

Intelligent inter-view coding method
Technical Field
The invention relates to the field of 3D video coding, in particular to an intelligent inter-view coding method.
Background
The large amount of 3D video data presents unprecedented difficulties and challenges to the storage and transmission of video. Therefore, how to realize efficient 3D video coding has important theoretical research significance and practical application value. Currently, 3D-HEVC (3D high efficiency video coding) developed by JCT-VC, an international standard organization for video coding consisting of MPEG (moving picture experts group) and VCEG (video coding experts group) jointly, is the latest 3D video coding standard. The 3D-HEVC introduces a new prediction technology and a new coding tool for non-independent viewpoints and depth videos, increases the technology suitable for multi-viewpoint video coding and depth video coding, and improves the coding efficiency. However, the inter-view prediction technique of multi-view video coding still has a great room for improvement.
Currently, people mainly focus on the acquisition of disparity vectors and disparity compensated prediction. Chen and the like calculate the disparity vector of the current block by utilizing the correlation between the disparity vector of the video space-time domain adjacent coding block and the disparity vector of the current block, and reduce the dependency of the disparity vector on the depth map coding. Zhang et al propose a method for obtaining disparity vector of a current coding block only from spatial domain and time domain neighboring blocks of the current coding block and performing inter-view prediction, which improves coding efficiency. Woontack et al propose an overlapped block disparity compensation and adaptive window stereo image coding scheme. Wong et al analyze the horizontal scaling and cropping phenomena existing between views in detail, and in order to implement disparity compensation prediction based on HSS (horizontal scaling and cropping) with minimum complexity, an efficient sub-sampling block matching technique is adopted, thereby effectively achieving bit saving of multi-view video coding.
Deep learning achieves great results on tasks such as processing classification and super-resolution reconstruction. With the above introduction, in recent years, researchers at home and abroad develop researches on planar color video coding based on deep learning, and the coding performance is remarkably improved in the aspects of intra-frame prediction, inter-frame prediction, post-filtering processing, end-to-end image coding and the like.
Cui et al propose a convolutional neural network for intra prediction that takes coded blocks around a predicted block as input, can sufficiently learn texture information of a reference block, and provides an accurate prediction result. Yan et al convert fractional pixel motion compensation to the problem of regression between images, solve this problem with convolutional neural networks, and improve coding efficiency. Liu et al propose a motion compensation enhancement algorithm based on CNN (convolutional neural network), which not only performs motion compensation on the current block, but also further improves the accuracy of prediction by using the neighboring reconstruction region of the current coding block. Li et al propose a frame-level dynamic metadata post-processing method, which first classifies video frames according to content complexity and quality, transmits classification flags to a decoding end as auxiliary information, and then performs image enhancement on the frames of different classes by using a 20-layer convolutional neural network, thereby improving filtering performance. However, there is little research on improving the 3D video coding performance by using the deep learning technique, and particularly, there is no intelligent inter-view coding method at present, so it is the fundamental idea of the present invention to propose an intelligent inter-view coding method.
In the process of implementing the invention, the inventor finds that at least the following disadvantages and shortcomings exist in the prior art:
the existing algorithm for modeling the parallax relation can perform operations such as stretching and shearing on a reference image, although the accuracy of prediction can be effectively improved, the applicability is poor, and the effect is poor for videos with complex textures and rich actions; at present, an intelligent inter-view coding method is lacked.
Disclosure of Invention
The invention provides an intelligent inter-viewpoint coding method, which designs an intelligent inter-viewpoint prediction model based on a convolutional neural network by fully mining the high similarity between adjacent viewpoints and utilizing prediction information in a 3D video coding process to obtain more accurate prediction of a current coding block, thereby further improving the coding efficiency, and the details are described as follows:
a method of intelligent inter-view coding, the method comprising the steps of:
respectively acquiring a first prediction block obtained based on a disparity vector mode of an adjacent block and a second prediction block obtained based on a traditional disparity compensation prediction mode;
constructing a residual learning convolution neural network structure with variable convolution kernel size; the result of the cascade connection of the brightness component channels of the two prediction blocks is used as the input of a convolutional neural network, and the output of the convolutional neural network is the enhanced prediction block of the original image block;
training four different network models according to four different quantization parameters;
respectively extracting brightness components of the first prediction block and the second prediction block, performing channel cascade, and calling corresponding network models according to the size of the current coding block and a quantization parameter value of a cascaded result;
and finally, calculating rate distortion cost of the enhanced prediction block, comparing the rate distortion cost with the optimal mode cost of the current coding block, selecting the method as the optimal coding mode if the rate distortion cost is less than the optimal mode cost, and coding and transmitting the corresponding mark to a decoding end.
The obtaining of the first prediction block obtained based on the neighboring block disparity vector mode specifically includes:
if the disparity vector obtained based on the disparity vector mode of the adjacent block exists, setting a mark to record the position of the disparity vector in a final candidate list, storing a prediction block obtained by compensating the disparity vector, and outputting a prediction image after a complete frame image is coded;
and marking the block which exists and is predicted by using the disparity vector as 1, otherwise, marking the block as 0, outputting the mark in a form of a table, cutting the predicted image, and only reserving the predicted block which uses the disparity vector mode of the adjacent block as training data.
Further, the obtaining of the second prediction block obtained based on the conventional disparity compensated prediction mode specifically includes:
finding out an inter-view reference picture by judging a view index of the reference picture, performing parallax compensation by using the reference picture, optimizing rate distortion, and storing an optimal prediction result in different PU partition modes as a DCP prediction block;
and outputting a predicted image after the complete frame image is coded, cutting the predicted image by using the table, and outputting and storing the cut predicted image in sequence.
Wherein, the constructing of the residual learning convolutional neural network with variable convolutional kernel size can be expressed as:
Figure BDA0002175028880000031
wherein f (-) is a prediction function,
Figure BDA0002175028880000032
representing a cascade of channels, l { -,. represents a loss function for calculating the error between the predicted value and the true value, the first prediction block being denoted as P1The second prediction block is denoted as P2The original image block is designated as Label and is denoted as Y.
The technical scheme provided by the invention has the beneficial effects that:
1. according to the method, inter-viewpoint prediction information is fully utilized, a depth learning technology is introduced into the field of 3D video inter-viewpoint prediction coding, and a prediction block which is more accurate is generated by utilizing the learning characteristic of a convolutional neural network and is closer to an original image block, so that the residual error between the original image block and the prediction block is reduced, the bit overhead required by coding transmission is reduced, and the coding efficiency is further improved;
2. the invention adopts a data driving mode and trains a convolutional neural network by utilizing abundant and diverse data, so that the invention has better generalization.
Drawings
Fig. 1 is a flowchart of an intelligent inter-view coding method.
Fig. 2 is a network architecture diagram of intelligent inter-view prediction.
Detailed Description
In order to make the objects, technical solutions and advantages of the present invention more apparent, embodiments of the present invention are described in further detail below.
The embodiment of the invention provides an intelligent inter-view coding method, which fuses a prediction block obtained in an NBDV (neighboring block disparity vector) mode and a prediction block obtained in a DCP (traditional disparity compensated prediction) mode in a coding process, and generates a more accurate prediction block by utilizing the learning characteristic of a convolutional neural network, and the specific implementation steps are as follows:
firstly, acquiring training data
The prediction based on NBDV and the DCP are two different inter-view prediction coding modes in 3D video coding, however, some blocks to be coded do not have NBDV vectors, and therefore, in the process of acquiring training data, a prediction block obtained by the NBDV process and a DCP prediction block corresponding to the prediction block need to be acquired respectively. The size of the prediction block obtained in the embodiment of the present invention is 64 × 64, and a smaller prediction block is subsequently tested.
1) Acquisition of NBDV data
The NBDV mode exists in a Merge mode, a disparity vector (marked as IvDC) obtained through an NBDV process can be used as a prediction candidate in a Merge candidate list, each candidate in the list can be traversed in an encoding process, at the moment, if the IvDC exists, an extra mark is set to record the position of the disparity vector in a final candidate list, and meanwhile, a prediction block obtained through compensation by using the IvDC vector is stored.
However, not all blocks to be coded can obtain disparity vectors by NBDV, nor all IvDC vectors can be stored in the final candidate list, so the block that exists and is predicted using IvDC is marked as 1, otherwise it is marked as 0. And outputting a prediction image after the complete frame image is coded, outputting the marks in a form of a table, finally cutting the output prediction result by using matlab in combination with the corresponding table, and outputting the prediction result in sequence, wherein only the prediction block using the NBDV mode is reserved as training data.
2) Acquisition of DCP data
Disparity compensated prediction and motion compensated prediction have similar concepts, and can be understood as a method of inter prediction. But the reference frames of the two are essentially different. The reference frames for Motion Compensated Prediction (MCP) are encoded frames from the same view at different time instances, while Disparity Compensated Prediction (DCP) refers to encoded frames from different views at the same time instance. Because the DCP and the MCP have the function of heteroworking, the DCP mode and the MCP mode are mixed in the inter-frame prediction mode. The reference picture list is traversed during the motion estimation process of inter-frame prediction/inter-view prediction, at this time, an inter-view reference picture is found by judging the view index of the reference picture, and a prediction block obtained by compensating the reference picture is stored.
In addition, considering that the Merge mode can only be applied to a prediction block with a PU size of 2N × 2N, for a coding unit CU of the same size, the inter prediction mode traverses all PU partition modes, including 2N × 2N, 2N × N, and N × N. Here, in order to obtain a prediction block with higher quality, the embodiment of the present invention employs a rate distortion optimization technique to compare prediction results in different PU partition modes, and stores an optimal prediction result in different PU partition modes as a DCP prediction block.
In order to find the DCP prediction block corresponding to the NBDV prediction block from the output prediction graph, the table obtained in the process of obtaining NBDV data is reused, and the table is cut and sequentially output and stored.
Second, construction of prediction network
The embodiment of the invention is based on a network structure of VRCNN (variable convolution kernel size residual learning convolution neural network), NBDV data and DCP data are obtained by using the method in the previous step, and a prediction block obtained through an NBDV mode is marked as P1The prediction block obtained by the DCP mode is denoted as P2And taking the corresponding original image block as a Label and recording the Label as Y. Through the network, a better prediction is obtained, so that the prediction block is as close to the original image as possible, and the formula is:
Figure BDA0002175028880000051
wherein f (-) is a prediction function,
Figure BDA0002175028880000054
representing the channel cascade, and l {. and · } representing a loss function for calculating an error between a predicted value and a true value. From a regression point of view, the formula can be written as follows:
Figure BDA0002175028880000052
wherein r (-) represents a convolutional neural network, θ is a network parameter, PiFor input data of the network, YiIs its corresponding Label. The loss function l {, · } can be specifically written as:
Figure BDA0002175028880000053
three, network model training
The embodiment of the invention takes the result of the concatenation of the NBDV prediction block and the DCP prediction block channel as the input of the network, and the output of the network is an enhanced prediction block which is closer to the original image block. The network comprises four convolution layers in total, wherein each layer except the last layer uses ReLU as a nonlinear activation function, and f (x) is max (0, x), the second layer and the third layer are divided into two branches, characteristics are extracted by convolution kernels with different sizes and are fused, and the fused characteristics are used as input of the next layer. And finally, linearly superposing the input two-channel data by adopting a residual error learning mode, and adding the linearly superposed two-channel data and the output of the last layer of the network to obtain a final enhanced prediction block, so that the convergence of the network is facilitated, and the accuracy of network prediction can be improved. The specific network structure is shown in fig. 2.
The embodiment of the invention trains on a Caffe platform, different models (25_34, 30_39, 35_42 and 40_45) are trained aiming at four different QPs, and the size of the trained block is 64 multiplied by 64. Different models, such as 32 × 32, 16 × 16, 8 × 8, are then trained on blocks of different sizes. The training and testing in the embodiments of the present invention are performed only for the luminance component.
Fourth, embedded coding frame
The coding platform used in the embodiment of the invention is HTM16.2, a network and a coding frame are integrated, firstly, a prediction block obtained in an NBDV mode and a brightness component of a corresponding DCP prediction block are respectively extracted, channel cascade is carried out, then, a corresponding network model is called according to the size of a current block and a QP (quantization parameter) value according to the cascade result, finally, rate distortion cost is calculated for the prediction block output by the network and compared with the optimal mode cost of the current coding block, if the rate distortion cost is less than the optimal mode cost, the method provided by the invention is selected as an optimal coding mode, and a corresponding mark is coded and transmitted to a decoding end.
Wherein, the mode with the minimum cost is taken as the optimal mode.
In the embodiment of the present invention, except for the specific description of the model of each device, the model of other devices is not limited, as long as the device can perform the above functions.
Those skilled in the art will appreciate that the drawings are only schematic illustrations of preferred embodiments, and the above-described embodiments of the present invention are merely provided for description and do not represent the merits of the embodiments.
The above description is only for the purpose of illustrating the preferred embodiments of the present invention and is not to be construed as limiting the invention, and any modifications, equivalents, improvements and the like that fall within the spirit and principle of the present invention are intended to be included therein.

Claims (4)

1. An intelligent inter-view coding method, comprising the steps of:
respectively acquiring a first prediction block obtained based on a disparity vector mode of an adjacent block and a second prediction block obtained based on a traditional disparity compensation prediction mode;
constructing a residual learning convolution neural network structure with variable convolution kernel size; the result of the cascade connection of the brightness component channels of the two prediction blocks is used as the input of a convolutional neural network, and the output of the convolutional neural network is the enhanced prediction block of the original image block;
training four different network models according to four different quantization parameters;
respectively extracting brightness components of the first prediction block and the second prediction block, performing channel cascade, and calling corresponding network models according to the size of the current coding block and a quantization parameter value of a cascaded result;
and finally, calculating rate distortion cost of the enhanced prediction block, comparing the rate distortion cost with the optimal mode cost of the current coding block, selecting the method as the optimal coding mode if the rate distortion cost is less than the optimal mode cost, and coding and transmitting the corresponding mark to a decoding end.
2. The method of claim 1, wherein the obtaining of the first prediction block obtained based on the neighboring block disparity vector pattern is specifically:
if the disparity vector obtained based on the disparity vector mode of the adjacent block exists, setting a mark to record the position of the disparity vector in a final candidate list, storing a prediction block obtained by compensating the disparity vector, and outputting a prediction image after a complete frame image is coded;
and marking the block which exists and is predicted by using the disparity vector as 1, otherwise, marking the block as 0, outputting the mark in a form of a table, cutting the predicted image, and only reserving the predicted block which uses the disparity vector mode of the adjacent block as training data.
3. The method according to claim 1, wherein the obtaining of the second prediction block obtained based on the conventional disparity compensated prediction mode specifically comprises:
finding an inter-view reference picture by judging a view index of the reference picture, performing parallax compensation by using the reference picture, optimizing rate distortion, and storing an optimal prediction result in different PU partition modes as a second prediction block obtained in a traditional parallax compensation prediction mode;
and outputting a predicted image after the complete frame image is coded, cutting the predicted image by using the table, and outputting and storing the cut predicted image in sequence.
4. The intelligent inter-view coding method according to claim 1, wherein the constructing of the residual learning convolutional neural network with variable convolutional kernel size is represented as:
Figure FDA0003278741920000021
wherein f (-) is a prediction function,
Figure FDA0003278741920000022
representing a cascade of channels, l { -,. represents a loss function for calculating the error between the predicted value and the true value, the first prediction block being denoted as P1The second prediction block is denoted as P2Original image block creationIs Label and is denoted as Y.
CN201910775875.1A 2019-08-21 2019-08-21 Intelligent inter-view coding method Active CN110557646B (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
CN201910775875.1A CN110557646B (en) 2019-08-21 2019-08-21 Intelligent inter-view coding method

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
CN201910775875.1A CN110557646B (en) 2019-08-21 2019-08-21 Intelligent inter-view coding method

Publications (2)

Publication Number Publication Date
CN110557646A CN110557646A (en) 2019-12-10
CN110557646B true CN110557646B (en) 2021-12-07

Family

ID=68737861

Family Applications (1)

Application Number Title Priority Date Filing Date
CN201910775875.1A Active CN110557646B (en) 2019-08-21 2019-08-21 Intelligent inter-view coding method

Country Status (1)

Country Link
CN (1) CN110557646B (en)

Families Citing this family (4)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN111464814B (en) * 2020-03-12 2022-01-04 天津大学 A virtual reference frame generation method based on parallax-guided fusion
CN113949867B (en) * 2020-07-16 2023-06-20 武汉Tcl集团工业研究院有限公司 Image processing method and device
CN116472707A (en) * 2020-09-30 2023-07-21 Oppo广东移动通信有限公司 Image prediction method, encoder, decoder, and computer storage medium
EP4240013A4 (en) * 2020-10-28 2023-09-06 Guangdong Oppo Mobile Telecommunications Corp., Ltd. IMAGE PREDICTION METHOD, ENCODER, DECODER AND COMPUTER STORAGE MEDIUM

Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104429076A (en) * 2012-07-10 2015-03-18 高通股份有限公司 Generalized residual prediction for scalable video coding and 3d video coding
CN107396124A (en) * 2017-08-29 2017-11-24 南京大学 Video-frequency compression method based on deep neural network

Family Cites Families (1)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
KR102535361B1 (en) * 2017-10-19 2023-05-24 삼성전자주식회사 Image encoder using machine learning and data processing method thereof

Patent Citations (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN104429076A (en) * 2012-07-10 2015-03-18 高通股份有限公司 Generalized residual prediction for scalable video coding and 3d video coding
CN107396124A (en) * 2017-08-29 2017-11-24 南京大学 Video-frequency compression method based on deep neural network

Non-Patent Citations (2)

* Cited by examiner, † Cited by third party
Title
Fast Mode Decision Using Inter-View and Inter-Component Correlations for Multiview Depth Video Coding;Jianjun Lei;《IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS》;20150831;第11卷(第4期);全文 *
基于网格的立体图像编码算法;杨嘉琛;《光电子·激光》;20100131;第21卷(第1期);全文 *

Also Published As

Publication number Publication date
CN110557646A (en) 2019-12-10

Similar Documents

Publication Publication Date Title
CN110557646B (en) Intelligent inter-view coding method
CN101867813B (en) Multi-view video coding method oriented for interactive application
CN102905150B (en) Novel multi-view video fractal coding, compressing and decompressing method
CN103338370B (en) A kind of multi-view depth video fast encoding method
CN103051894B (en) A kind of based on fractal and H.264 binocular tri-dimensional video compression & decompression method
CN101980538B (en) Fractal-based binocular stereoscopic video compression coding/decoding method
CN102291579B (en) A Fast Fractal Compression and Decompression Method for Multi-eye Stereo Video
CN101986716A (en) Quick depth video coding method
CN104469336B (en) Coding method for multi-view depth video signals
CN111464814B (en) A virtual reference frame generation method based on parallax-guided fusion
CN104768019A (en) A Neighboring Disparity Vector Acquisition Method for Multi-texture and Multi-depth Video
CN102045571B (en) Fast iterative search algorithm for stereo video coding
CN104038760A (en) Intra-frame wedge segmentation mode selection method and system for 3D (three-dimensional) video depth image
CN101198061A (en) Stereoscopic Video Stream Coding Method Based on View Image Mapping
CN103024381A (en) Macro block mode fast selecting method based on just noticeable distortion
CN111669601B (en) Intelligent multi-domain joint prediction coding method and device for 3D video
CN104539970A (en) 3D-HEVC interframe coding merge mode fast decision making method
CN102611893A (en) DMVC (distributed multi-view video coding) side-information integration method on basis of histogram matching and SAD (security association database) judgment
CN102420990B (en) Multi-view video-oriented fast coding method
CN103873867A (en) Free viewpoint video depth map distortion prediction method and free viewpoint video depth map coding method
CN102316323B (en) A Fast Fractal Compression and Decompression Method for Binocular Stereo Video
CN106210741B (en) A kind of deep video encryption algorithm based on correlation between viewpoint
CN101959067A (en) Fast Coding Mode Decision Method and System Based on Epipolar Constraints
Yan et al. CTU layer rate control algorithm in scene change video for free-viewpoint video
CN101980539B (en) Fractal-based multi-view three-dimensional video compression coding and decoding method

Legal Events

Date Code Title Description
PB01 Publication
PB01 Publication
SE01 Entry into force of request for substantive examination
SE01 Entry into force of request for substantive examination
GR01 Patent grant
GR01 Patent grant