空时形状预测与高效编码
Spatio-temporal shape prediction and efficient coding
- 2016年21卷第1期 页码:1-7
网络出版:2016-01-11,
纸质出版:2016
DOI: 10.11834/jig.20160101
移动端阅览

浏览全部资源
扫码关注微信
网络出版:2016-01-11,
纸质出版:2016
移动端阅览
形状是视觉对象的关键特征
形状编码是对象基图像和视频处理中的关键技术
但现有无损形状编码方法压缩效率普遍不高.为此
提出一种基于链码表示和空时预测的高效无损形状编码新算法. 首先逐帧提取视觉对象的形状轮廓并转化为链码表示;然后基于对象轮廓的帧间活动性将形状视频序列分成帧内预测编码帧和帧间预测编码帧
并基于轮廓链码的空域相关性和时域相关性对二者分别进行空域和时域补偿与预测;最后基于链码的方向约束特性对预测后的位移矢量和预测残差进行高效编码压缩. 为了检验所提算法的性能
基于MPEG-4标准形状测试序列进行了编码实验测试.与现有主要方法相比本文算法能提高压缩效率6%到71.6%不等. 本文算法可广泛应用于对象基编码、基于内容的图像检索、图像分析与理解等领域.
The use of a shape is a popular way to define objects
and efficient shape coding is a key technique in object-based applications. Shape coding is also a hot research topic in the field of image and video signal processing
and many shape-coding techniques have been proposed. Among these methods
chain-coding is a popular technique that can be used for lossless shape coding. However
most existing chain-based shape-coding methods have not exploited the spatio-temporal redundancy contained within shape image sequences. Similar to the existence of strong spatio-temporal redundancy within and among video textures
a strong redundancy also exists within and between object contours. This redundancy can be exploited to improve coding efficiency. Hence
in this paper
a novel chain-based lossless shape-coding scheme is proposed by exploiting the spatio-temporal correlations among object contours to acquire high coding efficiency. First
for a given shape image sequence
the contours of visual objects are extracted
thinned to perfect single-pixel width
and transformed into chain-based representation frame by frame. Second
the activity of object contours in each frame is detected and evaluated. The shape frames are classified into two coding categories on the basis of this activity: intra-coding frames and inter-coding frames. If the contour activity in a frame is larger than a preset threshold
the activity will be encoded as an inter-coding frame; otherwise
it will be encoded as an intra-coding frame. For an intra-coding frame
the spatial correlations within object contours are exploited on the basis of chain-based spatial prediction and compensation. For an inter-coding frame
the temporal correlations among object contours are exploited on the basis of chain-based temporal prediction and compensation. Finally
a new method is introduced to efficiently encode the prediction residuals and motion displacements by analyzing the constraints among chain links. To evaluate the performance of the proposed scheme
experiments are conducted and a partial comparison is performed against some well-known existing methods
including the lossless coding scheme proposed by the Joint Bi-level Image Experts Group (JBIG)
the improved lossless coding scheme proposed by JBIG (JBIG2)
the Context-based Arithmetic Encoding with Intra-mode (CAE Intra) of MPEG-4
the Context-based Arithmetic Encoding with Inter-mode (CAE Inter) of MPEG-4
the Digital Straight Line Segments-based Coding with Intra-mode ( DSLSC Intra) and the Digital Straight Line Segments-based Coding with Inter-mode (DSLSC Inter)
are also presented.
The experimental results show that the average code length of our scheme is only 28.4% of JBIG
32.3% of JBIG2
39.9% of CAE Intra
78.1% of CAE Inter
48.4% of DSLSC Intra
and 94.0% of DSLSC Inter. As a whole
the proposed scheme outperforms all existing techniques and is considerably more efficient than other methods. As far as we know
the DSLSC Inter is the most efficient lossless shape-coding approach. However
compared with the DSLSC Inter
the proposed scheme has an average code length that can be reduced by 6%. The proposed scheme has wide prospects in many object-based images and video applications
such as object-based coding
object-based editing
and object-based retrieval.
相关作者
相关机构
京公网安备11010802024621