Current Issue Cover
空时形状预测与高效编码

朱仲杰1, 王玉儿1, 蒋刚毅2(1.浙江万里学院宁波市DSP重点实验室, 宁波 315100;2.宁波大学电路与系统研究所, 宁波 315211)

摘 要
目的 形状是视觉对象的关键特征,形状编码是对象基图像和视频处理中的关键技术,但现有无损形状编码方法压缩效率普遍不高.为此,提出一种基于链码表示和空时预测的高效无损形状编码新算法.方法 首先逐帧提取视觉对象的形状轮廓并转化为链码表示;然后基于对象轮廓的帧间活动性将形状视频序列分成帧内预测编码帧和帧间预测编码帧,并基于轮廓链码的空域相关性和时域相关性对二者分别进行空域和时域补偿与预测;最后基于链码的方向约束特性对预测后的位移矢量和预测残差进行高效编码压缩.结果 为了检验所提算法的性能,基于MPEG-4标准形状测试序列进行了编码实验测试.与现有主要方法相比本文算法能提高压缩效率6%到71.6%不等.结论 本文算法可广泛应用于对象基编码、基于内容的图像检索、图像分析与理解等领域.
关键词
Spatio-temporal shape prediction and efficient coding

Zhu Zhongjie1, Wang Yuer1, Jiang Gangyi2(1.Ningbo Key Lab. of DSP, Zhejiang Wanli University, Ningbo 315100, China;2.Institute of circuits and system, Ningbo University, Ningbo 315211, China)

Abstract
Objective The use of a shape is a popular way to define objects, and efficient shape coding is a key technique in object-based applications. Shape coding is also a hot research topic in the field of image and video signal processing, and many shape-coding techniques have been proposed. Among these methods, chain-coding is a popular technique that can be used for lossless shape coding. However, most existing chain-based shape-coding methods have not exploited the spatio-temporal redundancy contained within shape image sequences. Similar to the existence of strong spatio-temporal redundancy within and among video textures, a strong redundancy also exists within and between object contours. This redundancy can be exploited to improve coding efficiency. Hence, in this paper, a novel chain-based lossless shape-coding scheme is proposed by exploiting the spatio-temporal correlations among object contours to acquire high coding efficiency. Method First, for a given shape image sequence, the contours of visual objects are extracted, thinned to perfect single-pixel width, and transformed into chain-based representation frame by frame. Second, the activity of object contours in each frame is detected and evaluated. The shape frames are classified into two coding categories on the basis of this activity: intra-coding frames and inter-coding frames. If the contour activity in a frame is larger than a preset threshold, the activity will be encoded as an inter-coding frame; otherwise, it will be encoded as an intra-coding frame. For an intra-coding frame, the spatial correlations within object contours are exploited on the basis of chain-based spatial prediction and compensation. For an inter-coding frame, the temporal correlations among object contours are exploited on the basis of chain-based temporal prediction and compensation. Finally, a new method is introduced to efficiently encode the prediction residuals and motion displacements by analyzing the constraints among chain links. Result To evaluate the performance of the proposed scheme, experiments are conducted and a partial comparison is performed against some well-known existing methods, including the lossless coding scheme proposed by the Joint Bi-level Image Experts Group (JBIG) , the improved lossless coding scheme proposed by JBIG (JBIG2), the Context-based Arithmetic Encoding with Intra-mode (CAE Intra) of MPEG-4, the Context-based Arithmetic Encoding with Inter-mode (CAE Inter) of MPEG-4, the Digital Straight Line Segments-based Coding with Intra-mode ( DSLSC Intra) and the Digital Straight Line Segments-based Coding with Inter-mode (DSLSC Inter), are also presented., The experimental results show that the average code length of our scheme is only 28.4% of JBIG, 32.3% of JBIG2, 39.9% of CAE Intra, 78.1% of CAE Inter, 48.4% of DSLSC Intra, and 94.0% of DSLSC Inter. Conclusion As a whole, the proposed scheme outperforms all existing techniques and is considerably more efficient than other methods. As far as we know, the DSLSC Inter is the most efficient lossless shape-coding approach. However, compared with the DSLSC Inter, the proposed scheme has an average code length that can be reduced by 6%. The proposed scheme has wide prospects in many object-based images and video applications, such as object-based coding, object-based editing, and object-based retrieval.
Keywords

订阅号|日报