首页 > 最新文献

IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.最新文献

英文 中文
Visual QoS programming environment for ubiquitous multimedia services 面向泛在多媒体业务的可视化QoS编程环境
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237785
Xiaohui Gu, D. Wichadakul, K. Nahrstedt
The provision of distributed multimedia services is becoming mobile and ubiquitous. Different multimedia services require application-specific Quality of Service (QoS). In this paper, we present QoSTalk, a unified component-based programming environment that allows application developers to specify different application-specific QoS requirements easily. In QoSTalk, we adopt a hierarchical approach to model application configuration graphs for different distributed multimedia services. We design and implement the XML-based Hierarchical QoS Markup Language, called HQML, to describe the hierarchical configuration graph as well as other application-specific QoS requirements and policies. QoSTalk promotes the separation of concerns in developing QoS-aware ubiquitous multimedia applications and thus enables easy programming of QoS-aware applications, running on top of a unified QoS-aware middleware framework. We have prototyped the QoSTalk in Java and CORBA. Our case studies with several multimedia applications show that QoSTalk effectively fills the gap for application developers between the very general facilities provided by the QoS-aware middleware and different kinds of distributed multimedia applications.
分布式多媒体服务的提供正变得移动化和无处不在。不同的多媒体业务需要特定于应用程序的服务质量(QoS)。在本文中,我们提出了QoSTalk,一个基于组件的统一编程环境,允许应用程序开发人员轻松地指定不同的特定于应用程序的QoS需求。在QoSTalk中,我们采用分层方法对不同分布式多媒体服务的应用程序配置图进行建模。我们设计并实现了基于xml的分层QoS标记语言,称为HQML,用于描述分层配置图以及其他特定于应用程序的QoS需求和策略。在开发支持qos的无处不在的多媒体应用程序时,QoSTalk促进了关注点的分离,从而使在统一的支持qos的中间件框架之上运行的支持qos的应用程序的编程变得容易。我们已经用Java和CORBA对QoSTalk进行了原型化。我们对几个多媒体应用程序的案例研究表明,QoSTalk有效地填补了应用程序开发人员在qos感知中间件和不同类型的分布式多媒体应用程序提供的非常通用的工具之间的空白。
{"title":"Visual QoS programming environment for ubiquitous multimedia services","authors":"Xiaohui Gu, D. Wichadakul, K. Nahrstedt","doi":"10.1109/ICME.2001.1237785","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237785","url":null,"abstract":"The provision of distributed multimedia services is becoming mobile and ubiquitous. Different multimedia services require application-specific Quality of Service (QoS). In this paper, we present QoSTalk, a unified component-based programming environment that allows application developers to specify different application-specific QoS requirements easily. In QoSTalk, we adopt a hierarchical approach to model application configuration graphs for different distributed multimedia services. We design and implement the XML-based Hierarchical QoS Markup Language, called HQML, to describe the hierarchical configuration graph as well as other application-specific QoS requirements and policies. QoSTalk promotes the separation of concerns in developing QoS-aware ubiquitous multimedia applications and thus enables easy programming of QoS-aware applications, running on top of a unified QoS-aware middleware framework. We have prototyped the QoSTalk in Java and CORBA. Our case studies with several multimedia applications show that QoSTalk effectively fills the gap for application developers between the very general facilities provided by the QoS-aware middleware and different kinds of distributed multimedia applications.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"42 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133601241","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 22
Lessons from speechreading 从演讲中得到的教训
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237780
P. Scanlon, R. Reilly
Speechreading is the ability to understand a speaker’s thoughts by watching the movements of the face and body and by using the information provided by the situation and the language. People with normal hearing and the hearing impaired use speechreading to augment communication especially in noisy environments. Just as people learn this skill, machines can be trained to understand a speakers meaning. Audio-Visual Automatic Speech Recognition (AV ASR) systems use audio and visual information to recognize what has been ‘said’. The speech sounds and movements provided need not be standard speech sounds or movements. The system will provide recognition given audio information only, visual information only or both.
读言是一种通过观察说话人面部和身体的动作,并利用情境和语言提供的信息来理解说话人思想的能力。听力正常的人和听力受损的人使用言语阅读来增强交流,特别是在嘈杂的环境中。就像人们学习这项技能一样,机器也可以被训练来理解说话人的意思。视听自动语音识别(AV ASR)系统使用音频和视觉信息来识别“所说”的内容。提供的语音和动作不必是标准的语音或动作。该系统将仅提供音频信息、仅提供视觉信息或两者兼而有之的识别。
{"title":"Lessons from speechreading","authors":"P. Scanlon, R. Reilly","doi":"10.1109/ICME.2001.1237780","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237780","url":null,"abstract":"Speechreading is the ability to understand a speaker’s thoughts by watching the movements of the face and body and by using the information provided by the situation and the language. People with normal hearing and the hearing impaired use speechreading to augment communication especially in noisy environments. Just as people learn this skill, machines can be trained to understand a speakers meaning. Audio-Visual Automatic Speech Recognition (AV ASR) systems use audio and visual information to recognize what has been ‘said’. The speech sounds and movements provided need not be standard speech sounds or movements. The system will provide recognition given audio information only, visual information only or both.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"61 4","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114021866","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Audio driven facial animation for audio-visual reality 音频驱动的面部动画的视听现实
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237848
T. Faruquie, Ashish Kapoor, Rohit J. Kate, Nitendra Rajput, L. V. Subramaniam
In this paper, we demonstrate a morphing based automated audio driven facial animation system. Based on an incoming audio stream, a face image is animated with full lip synchronization and expression. An animation sequence using optical flow between visemes is constructed, given an incoming audio stream and still pictures of a face speaking different visemes. Rules are formulated based on coarticulation and the duration of a viseme to control the continuity in terms of shape and extent of lip opening. In addition to this new viseme-expression combinations are synthesized to be able to generate animations with new facial expressions. Finally various applications of this system are discussed in the context of creating audio-visual reality.
在本文中,我们演示了一个基于变形的自动音频驱动面部动画系统。基于传入的音频流,人脸图像被动画化,具有完整的嘴唇同步和表情。在给定输入音频流和静止图像的情况下,使用不同视点之间的光流构建动画序列。规则是根据协发音和粘体的持续时间制定的,以控制形状和嘴唇张开程度的连续性。除此之外,新的面部表情组合被合成,能够生成新的面部表情动画。最后讨论了该系统在创建视听现实环境中的各种应用。
{"title":"Audio driven facial animation for audio-visual reality","authors":"T. Faruquie, Ashish Kapoor, Rohit J. Kate, Nitendra Rajput, L. V. Subramaniam","doi":"10.1109/ICME.2001.1237848","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237848","url":null,"abstract":"In this paper, we demonstrate a morphing based automated audio driven facial animation system. Based on an incoming audio stream, a face image is animated with full lip synchronization and expression. An animation sequence using optical flow between visemes is constructed, given an incoming audio stream and still pictures of a face speaking different visemes. Rules are formulated based on coarticulation and the duration of a viseme to control the continuity in terms of shape and extent of lip opening. In addition to this new viseme-expression combinations are synthesized to be able to generate animations with new facial expressions. Finally various applications of this system are discussed in the context of creating audio-visual reality.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"242 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116390316","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 13
Duration dependent input output markov models for audio-visual event detection 用于视听事件检测的时间依赖输入输出马尔可夫模型
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237704
M. Naphade, A. Garg, Thomas S. Huang
Detecting semantic events from audio-visual data with Spatiotemporal support is a challenging multimedia Understanding problem. The difficulty lies in the gap that exists between low level media features and high level semantic concept. We present a duration dependent input output Markov model (DDIOMM) to detect events based on multiple modalities. The DDIOMM combines the ability to model nonexponential duration densities with the mapping of input sequences to output sequences. In spirit it resembles the IOHMMs [1] as well as inhomogeneousHMMs [2]. We use the DDIOMM to model the audio-visual event explosion. We compare the detection performance of the DDIOMM with the IOMM as well as the HMM. Experiments reveal that modeling of duration improves detection performance.
在时空支持下从视听数据中检测语义事件是一个具有挑战性的多媒体理解问题。难点在于低层次的媒介特征与高层次的语义概念之间存在差距。我们提出了一个依赖于时间的输入输出马尔可夫模型(DDIOMM)来检测基于多模态的事件。DDIOMM将建模非指数持续时间密度的能力与输入序列到输出序列的映射相结合。在精神上,它类似于iohmm b[1]以及非均匀的iohmm b[2]。我们使用DDIOMM来模拟视听事件爆炸。我们比较了DDIOMM与IOMM以及HMM的检测性能。实验表明,持续时间建模可以提高检测性能。
{"title":"Duration dependent input output markov models for audio-visual event detection","authors":"M. Naphade, A. Garg, Thomas S. Huang","doi":"10.1109/ICME.2001.1237704","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237704","url":null,"abstract":"Detecting semantic events from audio-visual data with Spatiotemporal support is a challenging multimedia Understanding problem. The difficulty lies in the gap that exists between low level media features and high level semantic concept. We present a duration dependent input output Markov model (DDIOMM) to detect events based on multiple modalities. The DDIOMM combines the ability to model nonexponential duration densities with the mapping of input sequences to output sequences. In spirit it resembles the IOHMMs [1] as well as inhomogeneousHMMs [2]. We use the DDIOMM to model the audio-visual event explosion. We compare the detection performance of the DDIOMM with the IOMM as well as the HMM. Experiments reveal that modeling of duration improves detection performance.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116496677","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 17
An adaptive transmission scheme for audio and video synchronization based on real-time transport protocol 一种基于实时传输协议的音视频同步自适应传输方案
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237742
Chia-Chen Kuo, Ming-Syan Chen, Jeng-Chun Chen
Multimedia streams impose tight temporal constraints since different kinds of continuous multimedia streams have to be played synchronously. We devise in this paper an adaptive transmission scheme to ensure the continuous and synchronous playback of audio and video streams based on Real-time Transport Protocol. Realization of our adaptive scheme is composed of a series of operations in three stages, namely, (1) dynamic reordering mechanism, (2) decoding-recovery mechanism, and (3) adaptive synchronization mechanism. An empirical study is conducted to provide insights into our adaptive transmission scheme. As validated by our simulation results, the adaptive transmission mechanism is able to strike a good balance of both stable playback and the end-to-end delay reduction. Furthermore, we analyze the jitter resistance, the end-to-end delay, and the buffer size required in order to enhance the applicability of this scheme to more applications that require the transmission of multimedia data.
多媒体流施加了严格的时间限制,因为不同类型的连续多媒体流必须同步播放。本文设计了一种基于实时传输协议的自适应传输方案,以保证音频和视频流的连续同步播放。我们的自适应方案的实现由三个阶段的一系列操作组成,即(1)动态重排序机制,(2)解码恢复机制,(3)自适应同步机制。通过实证研究,对我们的自适应传输方案进行了深入研究。仿真结果表明,自适应传输机制能够在稳定回放和降低端到端延迟之间取得良好的平衡。此外,我们还分析了该方案的抗抖动性、端到端延迟和所需的缓冲区大小,以提高该方案在更多需要多媒体数据传输的应用中的适用性。
{"title":"An adaptive transmission scheme for audio and video synchronization based on real-time transport protocol","authors":"Chia-Chen Kuo, Ming-Syan Chen, Jeng-Chun Chen","doi":"10.1109/ICME.2001.1237742","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237742","url":null,"abstract":"Multimedia streams impose tight temporal constraints since different kinds of continuous multimedia streams have to be played synchronously. We devise in this paper an adaptive transmission scheme to ensure the continuous and synchronous playback of audio and video streams based on Real-time Transport Protocol. Realization of our adaptive scheme is composed of a series of operations in three stages, namely, (1) dynamic reordering mechanism, (2) decoding-recovery mechanism, and (3) adaptive synchronization mechanism. An empirical study is conducted to provide insights into our adaptive transmission scheme. As validated by our simulation results, the adaptive transmission mechanism is able to strike a good balance of both stable playback and the end-to-end delay reduction. Furthermore, we analyze the jitter resistance, the end-to-end delay, and the buffer size required in order to enhance the applicability of this scheme to more applications that require the transmission of multimedia data.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"116741556","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 15
A novel packet loss recovery technique for multimedia communication 一种新的多媒体通信丢包恢复技术
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237896
Wenqing Jiang, Antonio Ortega
In this paper a novel loss recovery technique is proposed for multimedia communications over lossy packet networks. The proposed technique uses a combination of recent results on multiple description coding and erasure recovery codes in channel coding. The uniqueness of the proposed technique lies in its ability to recover not only the data carried in lost packets, but also the decoding state for successive packets. Experimental results on image and speech coding show that the proposed technique has excellent coding performance compared to some of the best results published and it can also significantly reduce the error propagation in successive packets due to packet losses.
本文提出了一种基于有损分组网络的多媒体通信丢失恢复技术。该技术结合了信道编码中多重描述编码和擦除恢复编码的最新研究成果。该技术的独特之处在于它不仅能够恢复丢失数据包中携带的数据,而且能够恢复连续数据包的解码状态。图像和语音编码的实验结果表明,与已有的一些最佳编码结果相比,该技术具有优异的编码性能,并且可以显著减少由于丢包而导致的连续数据包中的错误传播。
{"title":"A novel packet loss recovery technique for multimedia communication","authors":"Wenqing Jiang, Antonio Ortega","doi":"10.1109/ICME.2001.1237896","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237896","url":null,"abstract":"In this paper a novel loss recovery technique is proposed for multimedia communications over lossy packet networks. The proposed technique uses a combination of recent results on multiple description coding and erasure recovery codes in channel coding. The uniqueness of the proposed technique lies in its ability to recover not only the data carried in lost packets, but also the decoding state for successive packets. Experimental results on image and speech coding show that the proposed technique has excellent coding performance compared to some of the best results published and it can also significantly reduce the error propagation in successive packets due to packet losses.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114664523","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
MPEG-2 multi-program transport stream transcoder MPEG-2多程序传输流转码器
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237747
Takeshi Takahashi, H. Kasai, T. Hanamura, H. Tominaga
MPEG-2 Multi-program Transport stream (TS) achieves improvement of transmission efficiency by multiplexing several MPEG-2 streams. In this paper, we propose a transcoder which achieves rate reduction of MPEG-2 multi-program TS. For the purpose of realizing MPEG-2 multi-program TS transcoder, this transcoder requires a rate control method and re-multiplexing method: The former improves average SNR values in total of streams, and the latter achieves the evasion from failure of STD buffer. Next, from simulation experiments, we compare the conventional rate control methods to the proposed one. On the other hand, we show the state of STD buffer. Finally, we show the effectiveness for our proposed scheme.
MPEG-2多程序传输流(Multi-program Transport stream, TS)通过将多个MPEG-2流复用来提高传输效率。本文提出了一种实现MPEG-2多程序TS降码的转码器,为了实现MPEG-2多程序TS转码器,该转码器需要一种速率控制方法和重复用方法,前者提高了总流的平均信噪比值,后者实现了对STD缓冲区失效的规避。然后,通过仿真实验,将传统的速率控制方法与本文提出的速率控制方法进行了比较。另一方面,我们显示了STD缓冲区的状态。最后,验证了该方案的有效性。
{"title":"MPEG-2 multi-program transport stream transcoder","authors":"Takeshi Takahashi, H. Kasai, T. Hanamura, H. Tominaga","doi":"10.1109/ICME.2001.1237747","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237747","url":null,"abstract":"MPEG-2 Multi-program Transport stream (TS) achieves improvement of transmission efficiency by multiplexing several MPEG-2 streams. In this paper, we propose a transcoder which achieves rate reduction of MPEG-2 multi-program TS. For the purpose of realizing MPEG-2 multi-program TS transcoder, this transcoder requires a rate control method and re-multiplexing method: The former improves average SNR values in total of streams, and the latter achieves the evasion from failure of STD buffer. Next, from simulation experiments, we compare the conventional rate control methods to the proposed one. On the other hand, we show the state of STD buffer. Finally, we show the effectiveness for our proposed scheme.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"82 2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115050116","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Using thesaurus to model keyblock-based image retrieval 使用同义词库建模基于键块的图像检索
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237683
Lei Zhu, Chun Tang, A. Rao, A. Zhang
Keyblock, which is a new framework we proposed for content-based image retrieval, is a generalization of the textbased information retrieval technology in the image domain. In this framework, keyblocks, which are analogous to keywords in text document retrieval, can be constructed by exploiting the method of Vector Quantization (VQ). Then an image can be represented as a list of keyblocks similar to a text document which can be considered as a list of keywords. Based on this image representation, various feature models can be constructed for supporting image retrieval. In this paper, we present a new feature representation model which use the keyblock-keyblock correlation matrix, termed keyblock-thesaurus, to facilitate the image retrieval. The feature vectors of this new model incorporate the effect of correlation between keyblocks, thus being more effective in representing image content.
Keyblock是我们提出的基于内容的图像检索框架,是基于文本的信息检索技术在图像领域的推广。在该框架中,可以利用向量量化(VQ)方法构造类似于文本文档检索中的关键字块。然后可以将图像表示为键块列表,类似于可以将文本文档视为关键字列表。基于这种图像表示,可以构建各种特征模型来支持图像检索。本文提出了一种新的特征表示模型,利用键块-键块相关矩阵(keyblock-thesaurus)来方便图像检索。该模型的特征向量考虑了键块之间的关联效应,能够更有效地表示图像内容。
{"title":"Using thesaurus to model keyblock-based image retrieval","authors":"Lei Zhu, Chun Tang, A. Rao, A. Zhang","doi":"10.1109/ICME.2001.1237683","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237683","url":null,"abstract":"Keyblock, which is a new framework we proposed for content-based image retrieval, is a generalization of the textbased information retrieval technology in the image domain. In this framework, keyblocks, which are analogous to keywords in text document retrieval, can be constructed by exploiting the method of Vector Quantization (VQ). Then an image can be represented as a list of keyblocks similar to a text document which can be considered as a list of keywords. Based on this image representation, various feature models can be constructed for supporting image retrieval. In this paper, we present a new feature representation model which use the keyblock-keyblock correlation matrix, termed keyblock-thesaurus, to facilitate the image retrieval. The feature vectors of this new model incorporate the effect of correlation between keyblocks, thus being more effective in representing image content.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131799330","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
Generation of personalized abstract of sports video 生成个性化的体育视频摘要
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237796
N. Babaguchi, Yoshihiko Kawai, T. Kitahashi
Video abstraction is defined as creating a shorter video clip from an original video stream. In this paper, we propose a method of generating a personalized abstract of broadcasted sports video. We first detect significant events from the video stream by matching with gamestats in which highlights of the game are described. Textual information in an overlay appearing on an image frame is recognized for this matching. Then, we select highlight shots from these detected events, reflecting on personal preferences. Finally, we connect each shot augmented with related audio and text in temporal order. From experimental results, we verified that an hourlength video can be compressed into a minute-length personalized abstract.
视频抽象定义为从原始视频流中创建更短的视频剪辑。本文提出了一种生成个性化体育转播视频摘要的方法。我们首先通过与描述游戏亮点的游戏统计相匹配,从视频流中检测重要事件。在图像帧上出现的覆盖层中的文本信息被识别为这种匹配。然后,我们根据个人喜好,从这些检测到的事件中选择高光镜头。最后,我们将每个镜头与相关的音频和文本按时间顺序连接起来。从实验结果中,我们验证了一小时长的视频可以压缩成一分钟长的个性化摘要。
{"title":"Generation of personalized abstract of sports video","authors":"N. Babaguchi, Yoshihiko Kawai, T. Kitahashi","doi":"10.1109/ICME.2001.1237796","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237796","url":null,"abstract":"Video abstraction is defined as creating a shorter video clip from an original video stream. In this paper, we propose a method of generating a personalized abstract of broadcasted sports video. We first detect significant events from the video stream by matching with gamestats in which highlights of the game are described. Textual information in an overlay appearing on an image frame is recognized for this matching. Then, we select highlight shots from these detected events, reflecting on personal preferences. Finally, we connect each shot augmented with related audio and text in temporal order. From experimental results, we verified that an hourlength video can be compressed into a minute-length personalized abstract.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129328416","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 34
Current status of WebCT and future of information basis for higher education 网络ct的现状与高等教育信息化基础的未来
Pub Date : 2001-08-22 DOI: 10.1109/ICME.2001.1237801
S. Kajita
WebCT that has been used widely in higher educational institutes of North America is going to be a mission critical e-Learning platform in on-campus education, rather than mere WBT system. In this paper, we describe WebCT and its current status in North America from the viewpoint of Japanese higher education. Then, we introduce three critical trends for educational information basis that we can observe in the movement of WebCT in North America; (1) contents exchange hub, (2) the integration of WebCT with existing student information system, and (3) campus portal that provides university-wide one-stop service for all member of the institution. Finally, we give a general view for educational information basis that should be contracted in Japanese higher educational institutes in 200X. We have already had all of technologies in our hands that would be necessary for higher educational institutions in the first decade of 21st century. How we can integrate and implement them to use in our own daily education seems to be a critical issue in the expected competitions in higher education.
网络ct已经在北美高等院校得到广泛应用,它将成为校园教育的关键任务电子学习平台,而不仅仅是WBT系统。本文从日本高等教育的角度来描述网络ct及其在北美的现状。然后,我们介绍了在北美WebCT运动中我们可以观察到的教育信息基础的三个关键趋势;(1)内容交换中心;(2)WebCT与现有学生信息系统的集成;(3)校园门户,为所有成员提供全校范围内的一站式服务。最后,对200X年日本高等院校应承包的教育信息基础提出了总体看法。我们已经掌握了21世纪头十年高等教育机构所需的所有技术。在高等教育的竞争中,如何将其整合并应用到我们的日常教育中,似乎是一个关键的问题。
{"title":"Current status of WebCT and future of information basis for higher education","authors":"S. Kajita","doi":"10.1109/ICME.2001.1237801","DOIUrl":"https://doi.org/10.1109/ICME.2001.1237801","url":null,"abstract":"WebCT that has been used widely in higher educational institutes of North America is going to be a mission critical e-Learning platform in on-campus education, rather than mere WBT system. In this paper, we describe WebCT and its current status in North America from the viewpoint of Japanese higher education. Then, we introduce three critical trends for educational information basis that we can observe in the movement of WebCT in North America; (1) contents exchange hub, (2) the integration of WebCT with existing student information system, and (3) campus portal that provides university-wide one-stop service for all member of the institution. Finally, we give a general view for educational information basis that should be contracted in Japanese higher educational institutes in 200X. We have already had all of technologies in our hands that would be necessary for higher educational institutions in the first decade of 21st century. How we can integrate and implement them to use in our own daily education seems to be a critical issue in the expected competitions in higher education.","PeriodicalId":405589,"journal":{"name":"IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2001-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122245206","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
期刊
IEEE International Conference on Multimedia and Expo, 2001. ICME 2001.
全部 Acc. Chem. Res. ACS Applied Bio Materials ACS Appl. Electron. Mater. ACS Appl. Energy Mater. ACS Appl. Mater. Interfaces ACS Appl. Nano Mater. ACS Appl. Polym. Mater. ACS BIOMATER-SCI ENG ACS Catal. ACS Cent. Sci. ACS Chem. Biol. ACS Chemical Health & Safety ACS Chem. Neurosci. ACS Comb. Sci. ACS Earth Space Chem. ACS Energy Lett. ACS Infect. Dis. ACS Macro Lett. ACS Mater. Lett. ACS Med. Chem. Lett. ACS Nano ACS Omega ACS Photonics ACS Sens. ACS Sustainable Chem. Eng. ACS Synth. Biol. Anal. Chem. BIOCHEMISTRY-US Bioconjugate Chem. BIOMACROMOLECULES Chem. Res. Toxicol. Chem. Rev. Chem. Mater. CRYST GROWTH DES ENERG FUEL Environ. Sci. Technol. Environ. Sci. Technol. Lett. Eur. J. Inorg. Chem. IND ENG CHEM RES Inorg. Chem. J. Agric. Food. Chem. J. Chem. Eng. Data J. Chem. Educ. J. Chem. Inf. Model. J. Chem. Theory Comput. J. Med. Chem. J. Nat. Prod. J PROTEOME RES J. Am. Chem. Soc. LANGMUIR MACROMOLECULES Mol. Pharmaceutics Nano Lett. Org. Lett. ORG PROCESS RES DEV ORGANOMETALLICS J. Org. Chem. J. Phys. Chem. J. Phys. Chem. A J. Phys. Chem. B J. Phys. Chem. C J. Phys. Chem. Lett. Analyst Anal. Methods Biomater. Sci. Catal. Sci. Technol. Chem. Commun. Chem. Soc. Rev. CHEM EDUC RES PRACT CRYSTENGCOMM Dalton Trans. Energy Environ. Sci. ENVIRON SCI-NANO ENVIRON SCI-PROC IMP ENVIRON SCI-WAT RES Faraday Discuss. Food Funct. Green Chem. Inorg. Chem. Front. Integr. Biol. J. Anal. At. Spectrom. J. Mater. Chem. A J. Mater. Chem. B J. Mater. Chem. C Lab Chip Mater. Chem. Front. Mater. Horiz. MEDCHEMCOMM Metallomics Mol. Biosyst. Mol. Syst. Des. Eng. Nanoscale Nanoscale Horiz. Nat. Prod. Rep. New J. Chem. Org. Biomol. Chem. Org. Chem. Front. PHOTOCH PHOTOBIO SCI PCCP Polym. Chem.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
0
微信
客服QQ
Book学术公众号 扫码关注我们
反馈
×
意见反馈
请填写您的意见或建议
请填写您的手机或邮箱
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
现在去查看 取消
×
提示
确定
Book学术官方微信
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术
文献互助 智能选刊 最新文献 互助须知 联系我们:info@booksci.cn
Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。
Copyright © 2023 Book学术 All rights reserved.
ghs 京公网安备 11010802042870号 京ICP备2023020795号-1