首页 > 最新文献

2008 IEEE 10th Workshop on Multimedia Signal Processing最新文献

英文 中文
Distortion estimation for optimized delivery of JPEG2000 compressed video with motion JPEG2000压缩视频优化传输中的失真估计
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665118
A. Naman, D. Taubman
A JPEG2000 compressed video sequence can provide better support for scalability, flexibility, and accessibility at a wider range of bit-rates than the current motion-compensated predictive video coding standards; however, it requires considerably more bandwidth to stream. The authors have recently proposed a novel approach that reduces the required bandwidth; this approach uses motion compensation and conditional replenishment of JPEG2000 code-blocks, aided by server-optimized selection of these code-blocks. The proposed approach can serve a diverse range of client requirements and can adapt immediately to interactive changes in client interests, such as forward or backward playback and zooming into individual frames. This work extends the previous work by approximating the distortion associated with the decisions made by the server without the need to recreate the actual video sequence at the server. The proposed distortion estimation algorithm is general and can be applied to various frames arrangements. Here, we choose to employ it in a hierarchical arrangement of frames, similar to the hierarchical B-frames of the SVC scalable video coding extension of the H.264/AVC standard. We employ a Lagrangian-style rate-distortion optimization procedure to the server transmission problem and compare the performance of both distortion estimation and exact distortion calculation cases against streaming individual frames and SVC. Results obtained suggest that the distortion estimation algorithm considerably reduces the amount of calculation needed by the server without enormously degrading the performance compared to the exact distortion calculation case. This work introduces the concepts, formulates the estimation and optimization problems, proposes a solution, and compares the performance to alternate strategies.
与当前的运动补偿预测视频编码标准相比,JPEG2000压缩视频序列可以在更大的比特率范围内提供更好的可扩展性、灵活性和可访问性支持;然而,它需要相当多的带宽来传输数据。作者最近提出了一种新颖的方法,可以减少所需的带宽;该方法使用JPEG2000代码块的运动补偿和条件补充,并借助于对这些代码块的服务器优化选择。所提出的方法可以满足各种各样的客户需求,并且可以立即适应客户兴趣的交互式变化,例如向前或向后播放以及缩放到单个帧。这项工作扩展了以前的工作,通过近似与服务器做出的决定相关的失真,而不需要在服务器上重新创建实际的视频序列。所提出的失真估计算法具有通用性,可适用于各种帧的排列。在这里,我们选择在帧的分层排列中使用它,类似于H.264/AVC标准的SVC可伸缩视频编码扩展的分层b帧。我们采用拉格朗日风格的速率失真优化程序来解决服务器传输问题,并比较了针对流式单独帧和SVC的失真估计和精确失真计算情况的性能。结果表明,与精确的失真计算情况相比,失真估计算法大大减少了服务器所需的计算量,而不会大大降低性能。本文介绍了概念,阐述了估计和优化问题,提出了解决方案,并将性能与备选策略进行了比较。
{"title":"Distortion estimation for optimized delivery of JPEG2000 compressed video with motion","authors":"A. Naman, D. Taubman","doi":"10.1109/MMSP.2008.4665118","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665118","url":null,"abstract":"A JPEG2000 compressed video sequence can provide better support for scalability, flexibility, and accessibility at a wider range of bit-rates than the current motion-compensated predictive video coding standards; however, it requires considerably more bandwidth to stream. The authors have recently proposed a novel approach that reduces the required bandwidth; this approach uses motion compensation and conditional replenishment of JPEG2000 code-blocks, aided by server-optimized selection of these code-blocks. The proposed approach can serve a diverse range of client requirements and can adapt immediately to interactive changes in client interests, such as forward or backward playback and zooming into individual frames. This work extends the previous work by approximating the distortion associated with the decisions made by the server without the need to recreate the actual video sequence at the server. The proposed distortion estimation algorithm is general and can be applied to various frames arrangements. Here, we choose to employ it in a hierarchical arrangement of frames, similar to the hierarchical B-frames of the SVC scalable video coding extension of the H.264/AVC standard. We employ a Lagrangian-style rate-distortion optimization procedure to the server transmission problem and compare the performance of both distortion estimation and exact distortion calculation cases against streaming individual frames and SVC. Results obtained suggest that the distortion estimation algorithm considerably reduces the amount of calculation needed by the server without enormously degrading the performance compared to the exact distortion calculation case. This work introduces the concepts, formulates the estimation and optimization problems, proposes a solution, and compares the performance to alternate strategies.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"18 4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130920332","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 6
Reversible watermarking of 3D mesh models by prediction-error expansion 基于预测误差扩展的三维网格模型可逆水印
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665183
Hao-Tian Wu, J. Dugelay
In this paper, a reversible watermarking algorithm is proposed for 3D mesh models based on prediction-error expansion. Firstly, we predict a vertex position by calculating the centroid of its traversed neighbors. Then the prediction error, i.e. the difference between the predicted and real positions, is expanded for data embedding. So only the vertex coordinates are modified to embed a watermark into the mesh content without changing the topology. We further reduce the distortion by adaptively choosing a threshold so that the prediction errors with too large magnitude will not be expanded. The chosen threshold value and critical location information should be saved in the watermarked mesh to guide the recovery process. The experiments show that the original mesh can be exactly recovered and consequently our algorithm can be used for symmetric or public key authentication of 3D mesh models.
提出了一种基于预测误差展开的三维网格模型可逆水印算法。首先,我们通过计算顶点经过的邻居的质心来预测顶点的位置。然后扩展预测误差,即预测位置与实际位置的差值,进行数据嵌入。因此,只需修改顶点坐标就可以在网格内容中嵌入水印,而无需改变拓扑结构。我们通过自适应选择阈值进一步减小了失真,使过大的预测误差不会扩大。选择的阈值和关键位置信息应保存在水印网格中,以指导恢复过程。实验表明,该算法可以准确地恢复原始网格,因此可以用于三维网格模型的对称或公钥认证。
{"title":"Reversible watermarking of 3D mesh models by prediction-error expansion","authors":"Hao-Tian Wu, J. Dugelay","doi":"10.1109/MMSP.2008.4665183","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665183","url":null,"abstract":"In this paper, a reversible watermarking algorithm is proposed for 3D mesh models based on prediction-error expansion. Firstly, we predict a vertex position by calculating the centroid of its traversed neighbors. Then the prediction error, i.e. the difference between the predicted and real positions, is expanded for data embedding. So only the vertex coordinates are modified to embed a watermark into the mesh content without changing the topology. We further reduce the distortion by adaptively choosing a threshold so that the prediction errors with too large magnitude will not be expanded. The chosen threshold value and critical location information should be saved in the watermarked mesh to guide the recovery process. The experiments show that the original mesh can be exactly recovered and consequently our algorithm can be used for symmetric or public key authentication of 3D mesh models.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"15 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134381887","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 43
Hybrid low bitrate audio coding using adaptive gain shape vector quantization 采用自适应增益形状矢量量化的混合低比特率音频编码
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665207
S. Mehrotra, Weig-Ge Chen, K. Koishida, Naveen Thumpudi
Audio coding at low bitrates typically suffers from artifacts caused by bandwidth truncation. In this paper we present a novel scheme to code audio signals at low bitrates which uses a traditional scalar quantization followed by entropy coding to code some portions of the spectrum (typically the lower portion). The other portions (typically the higher portions) of the spectrum are coded at a low bitrate using an adaptive gain shape vector quantizer where the codebook for vector quantization is formed by unmodified or modified versions of the portions of the spectrum which have already been coded. Fixed pre-trained codebooks are also available for use in certain cases. The use of such a scheme results in an audio codec which has been shown to be among the best audio codecs available at low bitrates. In addition, the decoder complexity of this audio codec is significantly lower than any other codec of equal quality at low bitrates.
低比特率的音频编码通常会受到带宽截断引起的伪影的影响。在本文中,我们提出了一种以低比特率编码音频信号的新方案,该方案使用传统的标量量化和熵编码来编码频谱的某些部分(通常是较低的部分)。频谱的其他部分(通常是较高的部分)使用自适应增益形状矢量量化器以低比特率编码,其中用于矢量量化的码本由已经编码的频谱部分的未修改或修改版本形成。固定的预训练密码本也可用于某些情况下。这种方案的使用产生的音频编解码器已被证明是在低比特率下可用的最佳音频编解码器之一。此外,这种音频编解码器的解码器复杂性在低比特率下明显低于任何其他同等质量的编解码器。
{"title":"Hybrid low bitrate audio coding using adaptive gain shape vector quantization","authors":"S. Mehrotra, Weig-Ge Chen, K. Koishida, Naveen Thumpudi","doi":"10.1109/MMSP.2008.4665207","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665207","url":null,"abstract":"Audio coding at low bitrates typically suffers from artifacts caused by bandwidth truncation. In this paper we present a novel scheme to code audio signals at low bitrates which uses a traditional scalar quantization followed by entropy coding to code some portions of the spectrum (typically the lower portion). The other portions (typically the higher portions) of the spectrum are coded at a low bitrate using an adaptive gain shape vector quantizer where the codebook for vector quantization is formed by unmodified or modified versions of the portions of the spectrum which have already been coded. Fixed pre-trained codebooks are also available for use in certain cases. The use of such a scheme results in an audio codec which has been shown to be among the best audio codecs available at low bitrates. In addition, the decoder complexity of this audio codec is significantly lower than any other codec of equal quality at low bitrates.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"24 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129462774","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Extraction of 3D scene structure for semantic annotation and retrieval of unedited video 三维场景结构的语义标注提取和未编辑视频的检索
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665053
I. Feldmann, W. Waizenegger, O. Schreer
In this paper we discuss the application of 3D scene reconstruction techniques in the area of automatic semantic annotation, search and retrieval of unedited video footage. Rather than working with static key-frames we exploit the time-depended dynamic properties of a moving camera. Based on state of the art camera self calibration techniques we develop a powerful analysis chain. We demonstrate, that the reconstructed 3D scene information can be used to generate both, accurate low level scene descriptors as well as meaningful medium and high level semantic information. We show, that the proposed algorithms work even in case of sparse data sets. The proposed algorithms provide a powerful working base for further investigations in the area of low, medium and high level extraction of semantic information for unedited video.
本文讨论了三维场景重建技术在自动语义标注、未编辑视频片段的搜索和检索等方面的应用。而不是与静态关键帧工作,我们利用时间依赖的动态属性的移动相机。基于最先进的相机自校准技术,我们开发了一个强大的分析链。我们证明,重建的三维场景信息既可以用于生成准确的低层场景描述符,也可以用于生成有意义的中高层语义信息。我们证明,即使在稀疏数据集的情况下,所提出的算法也有效。本文提出的算法为进一步研究低、中、高层次的未编辑视频语义信息提取提供了有力的工作基础。
{"title":"Extraction of 3D scene structure for semantic annotation and retrieval of unedited video","authors":"I. Feldmann, W. Waizenegger, O. Schreer","doi":"10.1109/MMSP.2008.4665053","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665053","url":null,"abstract":"In this paper we discuss the application of 3D scene reconstruction techniques in the area of automatic semantic annotation, search and retrieval of unedited video footage. Rather than working with static key-frames we exploit the time-depended dynamic properties of a moving camera. Based on state of the art camera self calibration techniques we develop a powerful analysis chain. We demonstrate, that the reconstructed 3D scene information can be used to generate both, accurate low level scene descriptors as well as meaningful medium and high level semantic information. We show, that the proposed algorithms work even in case of sparse data sets. The proposed algorithms provide a powerful working base for further investigations in the area of low, medium and high level extraction of semantic information for unedited video.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"5 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129836417","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
An optimized Multiple Description video codec for lossy packet networks 一种针对有损分组网络的优化的多描述视频编解码器
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665093
Moyuresh Biswas, M. Frater, J. Arnold, M. Pickering
The problem of resilient video transmission over lossy packet networks is addressed in this paper. We propose a rate-distortion optimized multiple description (MD) codec. Two different optimization controls of the codec are described that are suited to rates of packet loss, including the case where packets can travel over multiple paths through the network, with each path-dependent packet-loss probabilities. A packetization method optimized to work seamlessly with the proposed MD codec is also proposed. Simulations performed under various packet loss scenarios show the importance of the two optimizations and also that the proposed framework achieves significantly improved video quality when compared with similar techniques.
本文研究了在有损分组网络中弹性视频传输的问题。提出了一种率失真优化的多重描述(MD)编解码器。描述了适合于丢包率的两种不同的编解码器优化控制,包括数据包可以在网络中通过多条路径传输的情况,每个路径都依赖于丢包概率。本文还提出了一种优化的与所提出的MD编解码器无缝工作的分组方法。在各种丢包场景下进行的模拟表明了这两种优化的重要性,并且与类似技术相比,所提出的框架显著提高了视频质量。
{"title":"An optimized Multiple Description video codec for lossy packet networks","authors":"Moyuresh Biswas, M. Frater, J. Arnold, M. Pickering","doi":"10.1109/MMSP.2008.4665093","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665093","url":null,"abstract":"The problem of resilient video transmission over lossy packet networks is addressed in this paper. We propose a rate-distortion optimized multiple description (MD) codec. Two different optimization controls of the codec are described that are suited to rates of packet loss, including the case where packets can travel over multiple paths through the network, with each path-dependent packet-loss probabilities. A packetization method optimized to work seamlessly with the proposed MD codec is also proposed. Simulations performed under various packet loss scenarios show the importance of the two optimizations and also that the proposed framework achieves significantly improved video quality when compared with similar techniques.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133474625","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Highlight detection in soccer video using web-casting text 在足球视频使用网络广播文本高亮检测
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665148
Xiaole Ding, Yin-Jun Miao, Fan Bu, Lifeng Sun, Shiqiang Yang
Highlight detection is a challenge task in soccer video analysis. Using Web-casting text as external knowledge is proved to be a short cut to achieve both efficiency and effectiveness. Based on the previous framework using Web-casting text, we have improved the processes of video time detection and highlight boundary detection. Our method can detect the transparent time bar and can achieve acceptable precision in highlight boundary detection though the Web text time is not accurate at all. This progress can make the framework more robust in practice.
高光检测是足球视频分析中的一个难点。利用网播文本作为外部知识是一条既高效又有效的捷径。在之前使用web播文本的框架的基础上,我们改进了视频时间检测和突出显示边界检测的过程。我们的方法可以检测到透明的时间条,并且在网页文本时间不准确的情况下,可以达到可以接受的高亮边界检测精度。这一进展可以使该框架在实践中更加健壮。
{"title":"Highlight detection in soccer video using web-casting text","authors":"Xiaole Ding, Yin-Jun Miao, Fan Bu, Lifeng Sun, Shiqiang Yang","doi":"10.1109/MMSP.2008.4665148","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665148","url":null,"abstract":"Highlight detection is a challenge task in soccer video analysis. Using Web-casting text as external knowledge is proved to be a short cut to achieve both efficiency and effectiveness. Based on the previous framework using Web-casting text, we have improved the processes of video time detection and highlight boundary detection. Our method can detect the transparent time bar and can achieve acceptable precision in highlight boundary detection though the Web text time is not accurate at all. This progress can make the framework more robust in practice.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116288582","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Fast Johnson-Lindenstrauss Transform for robust and secure image hashing 快速约翰逊-林登施特劳斯变换鲁棒和安全的图像哈希
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665170
Xudong Lv, Z. J. Wang
Dimension reduction based techniques, such as singular value decomposition (SVD) and non-negative matrix factorization (NMF), have been proved to provide excellent performance for robust and secure image hashing by retaining the essential features of the original image matrix while preventing intentional attacks. In this paper, we introduce a recently proposed low-distortion, dimension reduction technique, referred as fast Johnson-Lindenstrauss transform (FJLT), and propose the use of FJLT for image hashing. FJLT shares the low-distortion characteristics of a random projection but requires a much lower complexity. These two desirable properties make it suitable for image hashing. Our experiment results show that the proposed FJLT-based hash yields good robustness under a wide range of attacks. Furthermore, the influence of secret key on the proposed hashing algorithm is evaluated by receiver operating characteristics (ROC) graph, revealing the efficiency of the proposed approach.
基于降维的技术,如奇异值分解(SVD)和非负矩阵分解(NMF),通过保留原始图像矩阵的基本特征,同时防止故意攻击,为鲁棒和安全的图像哈希提供了出色的性能。在本文中,我们介绍了一种最近提出的低失真,降维技术,称为快速约翰逊-林登施特劳斯变换(FJLT),并建议使用FJLT进行图像哈希。FJLT具有随机投影的低失真特性,但复杂度要低得多。这两个理想的属性使它适合图像散列。实验结果表明,提出的基于fjlt的哈希算法在各种攻击下具有良好的鲁棒性。此外,利用接收者工作特征(ROC)图评估密钥对所提哈希算法的影响,揭示了所提方法的有效性。
{"title":"Fast Johnson-Lindenstrauss Transform for robust and secure image hashing","authors":"Xudong Lv, Z. J. Wang","doi":"10.1109/MMSP.2008.4665170","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665170","url":null,"abstract":"Dimension reduction based techniques, such as singular value decomposition (SVD) and non-negative matrix factorization (NMF), have been proved to provide excellent performance for robust and secure image hashing by retaining the essential features of the original image matrix while preventing intentional attacks. In this paper, we introduce a recently proposed low-distortion, dimension reduction technique, referred as fast Johnson-Lindenstrauss transform (FJLT), and propose the use of FJLT for image hashing. FJLT shares the low-distortion characteristics of a random projection but requires a much lower complexity. These two desirable properties make it suitable for image hashing. Our experiment results show that the proposed FJLT-based hash yields good robustness under a wide range of attacks. Furthermore, the influence of secret key on the proposed hashing algorithm is evaluated by receiver operating characteristics (ROC) graph, revealing the efficiency of the proposed approach.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"12 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114988089","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Optimal LR-PET protection for scalable video streams over lossy channels with random delay 最佳的LR-PET保护可扩展的视频流在随机延迟的有损通道
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665119
Ruiqin Xiong, D. Taubman
This paper investigates the optimal PET protection for streaming scalably compressed streams over networks where the delivery time constraints allow limited retransmissions (LR) and the communication channels exhibit both random losses and delays. A key property must be considered in this scenario is the possibility that a packet successfully arrives at the receiver in time, even if its acknowledgment is not received by the sender at certain deadlines. This paper proposes an extended LRPET scheme, namely random-delay LR-PET, in which additional streams may be sent to provide supplemental protection for the packets whose acknowledgments are still missing at a specified time after the transmission. To determine the optimal protection in each transmission opportunity, hypotheses concerning the number of acknowledged packets and the effect of future retransmission are considered. As the key contribution of this paper, we develop a method to derive the effective overall recovery probability versus redundancy characteristic, which significantly simplifies the actual protection assignment procedure. This paper also demonstrates the benefits of the optimization strategy proposed for this random-delay LR-PET scheme and the cruciality of time selection for scheduling retransmission.
本文研究了网络上可伸缩压缩流的最佳PET保护,其中传输时间限制允许有限的重传(LR),通信信道显示随机损失和延迟。在此场景中必须考虑的一个关键属性是数据包及时成功到达接收方的可能性,即使发送方在特定期限内未收到其确认。本文提出了一种扩展的LRPET方案,即随机延迟LR-PET方案,该方案可以发送额外的流,对传输到指定时间后仍未收到确认的数据包提供补充保护。为了确定每个传输机会中的最佳保护,考虑了有关确认数据包数量和未来重传影响的假设。作为本文的关键贡献,我们开发了一种方法来推导有效的总体恢复概率与冗余特性,这大大简化了实际的保护分配过程。本文还论证了随机延迟LR-PET方案优化策略的优点,以及时间选择对重传调度的重要性。
{"title":"Optimal LR-PET protection for scalable video streams over lossy channels with random delay","authors":"Ruiqin Xiong, D. Taubman","doi":"10.1109/MMSP.2008.4665119","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665119","url":null,"abstract":"This paper investigates the optimal PET protection for streaming scalably compressed streams over networks where the delivery time constraints allow limited retransmissions (LR) and the communication channels exhibit both random losses and delays. A key property must be considered in this scenario is the possibility that a packet successfully arrives at the receiver in time, even if its acknowledgment is not received by the sender at certain deadlines. This paper proposes an extended LRPET scheme, namely random-delay LR-PET, in which additional streams may be sent to provide supplemental protection for the packets whose acknowledgments are still missing at a specified time after the transmission. To determine the optimal protection in each transmission opportunity, hypotheses concerning the number of acknowledged packets and the effect of future retransmission are considered. As the key contribution of this paper, we develop a method to derive the effective overall recovery probability versus redundancy characteristic, which significantly simplifies the actual protection assignment procedure. This paper also demonstrates the benefits of the optimization strategy proposed for this random-delay LR-PET scheme and the cruciality of time selection for scheduling retransmission.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"102 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128396404","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
A fast content-dependent interpolation approach via adaptive filtering 一种基于自适应滤波的快速内容相关插值方法
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665135
Hui Li, Yuhua Peng, W. Hwang
Improving the subjective quality and reducing the computational complexity of interpolation algorithms are important issues in video and network signal processing. To this end, we propose a fast adaptive image interpolation algorithm that classifies pixels and uses different linear interpolation kernels that are adaptive to the class of a pixel. Pixels are classified into regions relevant to the perception of an image, either in a texture region, an edge region, or a smooth region. Image interpolation is performed with Neville filters, which can be efficiently implemented by a lifting scheme. Since linear interpolation tends to over-smooth pixels in edge regions and texture regions, we apply the Laplacian operator to enhance the pixels in those regions. The results of simulations show that the proposed algorithm not only reduces the computational complexity of the process, but also improves the visual quality of the interpolated images.
提高插值算法的主观质量和降低插值算法的计算复杂度是视频和网络信号处理中的重要问题。为此,我们提出了一种快速自适应图像插值算法,该算法对像素进行分类,并使用不同的线性插值核来适应像素的类别。像素被分类到与图像感知相关的区域,无论是在纹理区域,边缘区域还是光滑区域。图像插值是用内维尔滤波器进行的,它可以通过提升方案有效地实现。由于线性插值会使边缘区域和纹理区域的像素过于光滑,我们使用拉普拉斯算子来增强这些区域的像素。仿真结果表明,该算法不仅降低了插值过程的计算复杂度,而且提高了插值图像的视觉质量。
{"title":"A fast content-dependent interpolation approach via adaptive filtering","authors":"Hui Li, Yuhua Peng, W. Hwang","doi":"10.1109/MMSP.2008.4665135","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665135","url":null,"abstract":"Improving the subjective quality and reducing the computational complexity of interpolation algorithms are important issues in video and network signal processing. To this end, we propose a fast adaptive image interpolation algorithm that classifies pixels and uses different linear interpolation kernels that are adaptive to the class of a pixel. Pixels are classified into regions relevant to the perception of an image, either in a texture region, an edge region, or a smooth region. Image interpolation is performed with Neville filters, which can be efficiently implemented by a lifting scheme. Since linear interpolation tends to over-smooth pixels in edge regions and texture regions, we apply the Laplacian operator to enhance the pixels in those regions. The results of simulations show that the proposed algorithm not only reduces the computational complexity of the process, but also improves the visual quality of the interpolated images.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"262 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133726483","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Effective video event detection via subspace projection 基于子空间投影的有效视频事件检测
Pub Date : 2008-11-05 DOI: 10.1109/MMSP.2008.4665043
Jialie Shen, D. Tao, Xuelong Li
This paper describes a new video event detection framework based on subspace selection technique. With the approach, feature vectors presenting different kinds of video information can be easily projected from different modalities onto an unified subspace, on which recognition process can be performed. The approach is capable of discriminating different classes and preserving the intra-modal geometry of samples within an identical class. Distinguished from the existing multi-modal detection methods, the new system works well when some modalities are not available. Experimental results based on soccer video and TRECVID news video collections demonstrate the effectiveness, efficiency and robustness of the proposed method for individual recognition tasks in comparison to the existing approaches.
本文提出了一种新的基于子空间选择技术的视频事件检测框架。利用该方法,可以很容易地将表示不同类型视频信息的特征向量从不同的模态投射到统一的子空间中,从而可以在该子空间上进行识别。该方法能够区分不同的类别,并保留同一类别内样本的模态几何形状。与现有的多模态检测方法不同,新系统在某些模态不可用的情况下也能很好地工作。基于足球视频和TRECVID新闻视频集的实验结果表明,与现有方法相比,该方法在单个识别任务中具有有效性、高效性和鲁棒性。
{"title":"Effective video event detection via subspace projection","authors":"Jialie Shen, D. Tao, Xuelong Li","doi":"10.1109/MMSP.2008.4665043","DOIUrl":"https://doi.org/10.1109/MMSP.2008.4665043","url":null,"abstract":"This paper describes a new video event detection framework based on subspace selection technique. With the approach, feature vectors presenting different kinds of video information can be easily projected from different modalities onto an unified subspace, on which recognition process can be performed. The approach is capable of discriminating different classes and preserving the intra-modal geometry of samples within an identical class. Distinguished from the existing multi-modal detection methods, the new system works well when some modalities are not available. Experimental results based on soccer video and TRECVID news video collections demonstrate the effectiveness, efficiency and robustness of the proposed method for individual recognition tasks in comparison to the existing approaches.","PeriodicalId":402287,"journal":{"name":"2008 IEEE 10th Workshop on Multimedia Signal Processing","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2008-11-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134079787","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
2008 IEEE 10th Workshop on Multimedia Signal Processing
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1