博碩士論文 945401005 詳細資訊




以作者查詢圖書館館藏 以作者查詢臺灣博碩士 以作者查詢全國書目 勘誤回報 、線上人數:76 、訪客IP:18.226.133.0
姓名 方志倫(Chih-Lun Fang)  查詢紙本館藏   畢業系所 電機工程學系
論文名稱 電腦視覺特徵值萃取於字幕視訊處理及視訊防手震系統設計之研究
(Design of Feature Extraction for Text-Video Processing and Video Stabilization System with Computer Vision-Based Techniques)
相關論文
★ 即時的SIFT特徵點擷取之低記憶體硬體設計★ 即時的人臉偵測與人臉辨識之門禁系統
★ 具即時自動跟隨功能之自走車★ 應用於多導程心電訊號之無損壓縮演算法與實現
★ 離線自定義語音語者喚醒詞系統與嵌入式開發實現★ 晶圓圖缺陷分類與嵌入式系統實現
★ 語音密集連接卷積網路應用於小尺寸關鍵詞偵測★ G2LGAN: 對不平衡資料集進行資料擴增應用於晶圓圖缺陷分類
★ 補償無乘法數位濾波器有限精準度之演算法設計技巧★ 可規劃式維特比解碼器之設計與實現
★ 以擴展基本角度CORDIC為基礎之低成本向量旋轉器矽智產設計★ JPEG2000靜態影像編碼系統之分析與架構設計
★ 適用於通訊系統之低功率渦輪碼解碼器★ 應用於多媒體通訊之平台式設計
★ 適用MPEG 編碼器之數位浮水印系統設計與實現★ 適用於視訊錯誤隱藏之演算法開發及其資料重複使用考量
檔案 [Endnote RIS 格式]    [Bibtex 格式]    [相關文章]   [文章引用]   [完整記錄]   [館藏目錄]   [檢視]  [下載]
  1. 本電子論文使用權限為同意立即開放。
  2. 已達開放權限電子全文僅授權使用者為學術研究之目的,進行個人非營利性質之檢索、閱讀、列印。
  3. 請遵守中華民國著作權法之相關規定,切勿任意重製、散佈、改作、轉貼、播送,以免觸法。

摘要(中) 近年來,電腦視覺已成為一項重要的研究領域,相關研究提出許多電腦視覺演算法,而這些演算法可用來設計各種不同的應用系統,這些應用系統可進一步於硬體平台上實現。基於此背景,這篇論文即以電腦視覺為基礎設計一個新穎的系統。而在電腦視覺中,特徵萃取是一項十分重要的技術,該技術可實現各種不同的系統。在眾多的特徵中,考量字幕特徵在視訊內容的語意特性以及全域運動特徵在防手震、視訊編碼的重要性,本論文針對字幕及全域運動特徵萃取與應用於字幕視訊處理及防手震進行系統層面的探討研究,
  對於字幕特徵萃取而言,現今有許多字幕被嵌入在視訊中,這些字幕有時是廣告、沒有用途的,因此,需要將字幕去除並修補視訊內容。然而,由於受到大尺寸字體、結構區域及各種不同類型視訊的影響,造成極少傳統方法能完整的修補整體視訊內容。為了回應此需求,這個研究提出了字幕視訊修補技術,這項技術是基於結構修復及紋理傳遞。為了修復結構區域,結構內插法使用新提出的旋轉區塊比對法來估算修補區域的初始位置,之後再更新修補區域的位置,周圍畫面的資訊用來填入結構區域。結構延伸法採用SP-Line曲線估算來修補結構區域,以達到不需手動修補的目標,最後衍生傳遞實現了紋理區域的修補。實驗操作採用真實的電視視訊,由實驗結果可得知所有的字幕區域都能被完整修補且保持空間時間域的一致性。此外,比較結果顯示本論文所提出方法的效能優於其它傳統方法。本論文方法的優點包含透過整合多個畫面的結構內容來減少設計的複雜度、及達到真實視訊結構內容修補的一致性。
  此外,本論文利用嵌入字幕特徵來達到智慧型多媒體的顯示模式,我們設計了一個字幕子母畫面顯示系統,這個系統可以擷取子頻道的字幕並且與主頻道的視訊結合,這個系統建置於雙核心平台以達到即時字幕擷取及顯示。這個研究提出一個排程方式以進行擷取及顯示的分工運作。其次,本論文設計一個資料傳輸機制用來達到有效的資料傳輸,當中有些資料可重複被利用。再者建立SIMD機制以加速字幕擷取中大量迴旋及累加的運算。針對最佳化標籤及填入工作,開發了四倍緩衝、多重記憶體及多重工作技術。從評估結果驗證了所提出的技術能加快字幕子母畫面擷取及顯示的處理速度,比較結果也顯示了這個機制在實現字幕擷取上比其它方法更有效率。
  對於全域運動特徵萃取而言,為了能移除視訊中不必要的晃動,本論文提出一個基於全域運動特徵萃取的視訊防手震演算法。為了達到即時視訊防手震,所提出的演算法實現於雙核心硬體平台。本論文之方法經由計算區域運動得知全域運動特徵值,而區域運動透過特徵區塊比對以減少運算量。基於靜止背景代表全域運動特徵值的假設,提出一個背景運動模型。首先,直方圖統計估算初始全域運動,之後,修正程序透過背景運動模型修正全域運動,最後,以所估算的全域運動對視訊內容進行防手震處理。此外,為了提升防手震的效能,提出最佳化相關技術。防手震工作被分割及排程於雙核心平台執行。函式簡化法對特徵點選擇中的回應函式進行最佳化處理。此外,以區域範圍記憶體存取及最佳化絕對誤差和提升特徵區塊比對的速度,同時也針對全域運動特徵值估算進行最佳化。實驗結果顯示所提出的防手震演算法能正確估算全域運動並產生正確的防手震視訊,比較結果也證實本論文所提出的方法有較其它方法更高的防手震效能。同時,基於驗證結果,所提出的最佳化方法能增加防手震效能以達到即時處理運算。
摘要(英) Computer vision has become an important research field in recent years. Many computer vision-based algorithms are proposed to design various novel systems. Theses systems could be further implemented and realized on an embedded platform. Based on this background, this thesis designs a novel computer vision-based system. Generally, feature extraction is an important technique in computer vision, which can realize various systems. Among the different features, text feature has high level of semantics, and global motion has more importance in video stabilization and video coding. Thus, this thesis addresses the feature extraction of text and global motion and its applications on text-video inpainting and video stabilization.
 For text feature extraction, today, more superimposed text is embedded within videos. Usually some text is unnecessary. Thus, one requires an approach to remove the text and inpaint the video. However, few conventional approaches inpaints the video well due to the large-sized text, structure regions, and various types of videos. In response, this study designed a text-video inpainting algorithm that poses text-video inpainting as structure repair and texture propagation. To repair the structure regions, the structure interpolation uses the new model’s rotated block matching to estimate the initial location of inpainted regions and later refine the coordinates of inpainted regions. The information in the neighboring frames then fills the structure regions. To inpaint the structure regions without tedious manual interaction, the structure extension utilizes the spline curve estimation. Afterwards, derivative propagation realizes the texture region inpainting. The experiment results are based on several real text-video, where all of the text regions were inpainted with spatio-temporal consistency. Additionally, comparisons present that the performance of the proposed algorithm is superior to those of conventional approaches. Its advantages include the reduction of design complexity by only integrating the structure information in multi-frame and the demonstration of structure consistency for realistic videos.
 Additionally, some text feature is important information. Thus, this research utilizes embedded text to achieve an intelligent multimedia display. We design a text in picture (TiP) display system which can extract the texts in the subchannel and then combine these texts with the main channel. This system was constructed on a dual-core platform to reach real-time text extraction and display. A schedulable design framework was proposed to partition the TiP display with text extraction in pipeline running. A data-aware transfer scheme was designed in which some data can be reused. Single instruction multiple data (SIMD) based mechanisms were created to enhance the computational efficiency on numerous convolutions and accumulations in text extraction. Quadruple buffering was manipulated to process the input/output in text extraction simultaneously. To optimize the labeling and filling tasks, the multi-banking and multi-tasking were developed. The evaluation results indicated that the proposed techniques can speed up the processing time of TiP display with text extraction. The equivalent comparison presented that the proposed techniques are more proficient at realizing text extraction.
 For global motion feature extraction, to remove the unwanted vibration in video, a robust video stabilization algorithm based on global motion feature extraction is proposed. To achieve real-time video stabilization, the proposed algorithm is realized on a dual-core embedded platform. In our approach, the global motion is calculated from the local motion. The local motion is derived from feature-centered block matching with lower computation. Based on the assumption that the motion of static background represents the global motion, a background motion model is proposed. The histogram-based computation operates the local motion for initial global motion estimation. Afterwards, global motion is refined by an updating procedure. The updating procedure updates the global motion based on the background motion model. Finally, the video is smoothed and stabilized based on the computed global motion. In addition, to enhance the performance of video stabilization on an embedded platform, several novel optimization approaches are proposed. The video stabilization tasks are partitioned and scheduled on the dual cores. A function simplification approach is designed to optimize the response function in feature-point selection task. Moreover, the speed of feature-centered block matching is enhanced by the region-based memory access and sum of absolute difference (SAD) optimization. As well as, the global motion estimation is optimized. The experimental results present that the proposed video stabilization approach can accurately estimate the global motion and produce well stabilized videos. The comparison also demonstrates our superior performance on video stabilization. Based on the evaluation results, the proposed optimization approaches can significantly increase the performance of video stabilization for real-time processing.
關鍵字(中) ★ 視訊防手震
★ 視訊修補
★ 字幕移除
★ 雙核心平台
★ 字幕擷取
★ 即時處理
★ 全域運動估算
關鍵字(英) ★ Video Stabilization
★ Text Extraction
★ Text Removal
★ Video Inpainting
★ Real-Time Processing
★ Global Motion Estimation
★ Dual-Core Embedded Platform
論文目次 Abstract I
List of Figures X
List of Tables XIV
Chapter 1 Introduction
1.1 Background 1
1.2 Motivation 4
1.2.1 Text-Video Processing 5
1.2.2 Video Stabilization 8
1.3 Overview of Text-Video Processing 8
1.3.1 Text-Video Inpainting 8
1.3.2 TiP Display 9
1.4 Overview of Video Stabilization System 11
1.5 Thesis Organization 13
Chapter 2 Design of Text-Video Inpainting
2.1 Background Study of Text-Video Inpainting 14
2.1.1 Image Inpainting 14
2.1.1.1 Pixel Domain based Approach 15
2.1.1.2 Image Decomposition 16
2.1.1.3 Frequency Domain based Approach 17
2.1.2 Video Inpainting 17
2.1.2.1 Traditional Block Matching 17
2.1.2.2 Spatio-Temporal Block Searching 18
2.1.2.3 Object-Background Inpainting 19
2.2 Text-Video Inpainting Using Structure Repair and Texture Propagation 20
2.2.1 Overview of Proposed Text-Video Inpainting Algorithm 20
2.2.1.1 Evaluation of Text-Video Inpainting Design 21
2.2.1.2 Proposed Algorithm Overview 22
2.2.2 Proposed Techniques for Text-Video Inpainting 24
2.2.2.1 Finding Similar Structure by Rotated Block Matching 26
2.2.2.2 Structure Repair in Spatio-Temporal Domain 30
2.2.2.3 Texture Inpainting with Derivative Propagation 35
2.3 Implementation Results 38
Chapter 3 Design of Text in Picture Display System
3.1 Background Study of Text in Picture Display System 50
3.1.1 Algorithm analysis 50
3.1.1.1 Text Detection 50
3.1.1.2 Text Localization 51
3.1.1.3 Text Extraction 52
3.1.2 Platform-Based Design 53
3.1.2.1 Dual-Core Platform Analysis 54
3.1.2.2 Optimization Techniques 55
3.2 Design of Text in Picture Display System 59
3.2.1 Text in Picture Display System Architecture 59
3.2.2 Proposed Techniques for Designing Text in Picture Display with Text Extraction on a Dual-Core Platform… 61
3.2.2.1 Utilization Enhancement by Schedulable Design Framework 62
3.2.2.2 Optimization of Data Transfer by Data-Aware Transfer Scheme 66
3.2.2.3 Operation Reduction by SIMD Based Mechanism 68
3.2.2.4 Task Optimization by Quadruple Buffering and Multiple Processing 71
3.3 Implementation Results 75
Chapter 4 Design of Video Stabilization
4.1 Background Study Video Stabilization 81
4.1.1 Algorithm Analysis 81
4.1.1.1 Feature Point-Based Video Stabilization 81
4.1.1.2 Optical Flow-Based Video Stabilization 83
4.1.1.3 Block Matching-Based Video Stabilization 83
4.1.2 Platform-Based Design Analysis 84
4.2 A Video Stabilization System with Background Motion Estimation and Smoothing for Digital Camera 85
4.2.1 Proposed Robust Video Stabilization Algorithm for Digital Camera 85
4.2.1.1 Overview of Proposed Video Stabilization Algorithm 85
4.2.1.2 Proposed Algorithm for Video Stabilization 86
4.2.2 Platform-Based Design of Proposed Video Stabilization Algorithm 94
4.2.2.1 Partition of Video Stabilization Tasks 94
4.2.2.2 Optimization of Feature Point Selection 96
4.2.2.3 Optimization of Feature-Centered Block Matching 98
4.2.2.4 Optimization of Global Motion Estimation 100
4.3 Implementation Results 101
4.3.1 Evaluation of Video Stabilization Results 101
4.3.2 Evaluation of Video Stabilization Optimization Performance 111
Chapter 5 Conclusions
5.1 Text-Video Inpainting 115
5.2 Text in Picture Display System 116
5.3 Video Stabilization 117
Reference 118
參考文獻 [1] M. R. Lyu, J. Song, and M. Cai, “A comprehensive method for multilingual video text detection, localization, and extraction,” IEEE Trans. Circuits Syst. Video Technol., vol. 15, no. 2, Feb. 2005.
[2] T. H. Tsai, Y. C. Chen, and C. L. Fang, “2DVTE: a two-directional text extractor for rapid and elaborate design,” Pattern Recognition, vol. 42, no. 7, Jul. 2009.
[3] W. Kim and C. Kim, “A new approach for overlay text detection and extraction from complex video scene,” IEEE Trans. on Image Processing, vol. 18, no. 2, Feb. 2009.
[4] C. Dorai, R. Bolle, N. Dimitrova, L. and Agnihotri, “MPEG-7 text description scheme,” ISO/MPEG M5206, MPEG Melbourne Meeting, Oct. 1999.
[5] C. W. Yi, C. M. Su, W. T. Chai, J. L. Huang, and T. C. Chiang, “G-Constellations: G-Sensor Motion Tracking Systems,” in Proc. of the IEEE 71st Vehicular Technology Conf., 2010.
[6] G. Bleser and D. Stricker, “Advanced tracking through efficient image processing and visual-inertial sensor fusion,” in Proc. of the IEEE Virtual Reality Conf., 2008.
[7] C. N. Chiu, C. T. Tseng, and C. J. Tsai, “Tightly-coupled MPEG-4 video encoder framework on asymmetric dual-core platforms,” in Proc. IEEE Int'l Symposium on Circuits and Syst., 2005.
[8] K. Rapaka, M. Mody, and K. Prasad, “Scalable software architecture for high performance video codec's on parallel processing engines,” in Proc. IEEE Int'l Symposium on Consumer Electronics, 2007.
[9] K. V. Nedovodeev, “Multimedia data processing on dual-core SoC multicore-24,” in Proc. IEEE Int'l Symposium on Consumer Electronics, 2006.
[10] C. H. Yen, Y. S. Lin, and B. F. Wu, “An efficient implementation of a low-complexity MP3 algorithm with a stream cipher,” Multimedia Tools and Applications, vol. 35, no. 3, Dec. 2007.
[11] M. Bertalmio, G. Sapiro, V. Caselles, and C. Ballester, “Image inpainting,” in Proc. ACM Conf. Comp. Graphics (SIGGRAPH), Jul. 2000.
[12] M. Bertalmio, A. L. Bertozzi, and G. Sapiro, “Navier-stokes, fluid dynamics, and image and video inpainting,” in Proc. of the IEEE Int'l Conf. on Computer Vision and Pattern Recognition, pp. 355-362, 2001.
[13] C. Ballester, M. Bertalmio, V. Caselles, G. Sapiro, and J. Verdera, “Filling-in by joint interpolation of vector fields and gray levels,” IEEE Trans. Image Processing, vol. 10, no. 8, pp. 1200-1211, Aug. 2001.
[14] M. M. Oliveira, B. Bowen, R. McKenna, and Y. S. Chang, “Fast digital image inpainting,” in Proc. Conf. on Visualization, Imaging and Image Processing, pp. 261-266, 2001.
[15] A. Criminisi, P. Perez, and K. Toyama, “Region filling and object removal by exemplar-based image inpainting,” IEEE Trans. Image Processing, vol. 13, no. 9, pp. 1200-1212, Sep. 2004.
[16] N. Komodakis and G. Tziritas, “Image completion using global optimization,” in Proc. of the IEEE Int'l Conf. on Computer Vision and Pattern Recognition, pp. 442-452, 2006.
[17] T. K. Shih and R. C. Chang, “Digital inpainting - survey and multilayer image inpainting algorithms,” in Proc. of the third Int'l Conf. on Information Technology and Application, pp. 15-24, 2005.
[18] J. Sun, L. Yuan, J. Jia, and H. Y. Shum, “Image completion with structure propagation,” in Proc. ACM Conf. Comp. Graphics (SIGGRAPH), pp. 861-868, 2005.
[19] M. Bertalmio, V. Caselles, G. Sapiro, and S. Osher, “Simultaneous structure and texture image inpainting,” IEEE Trans. Image Processing, vol. 12, no. 8, pp. 882-889, Aug. 2003.
[20] Y. L. Chen, C. T. Hsieh, and C. H. Hsu, “Progressive image inpainting based on wavelet transform,” IEICE Trans.Fundamentals, vol. E88-A, no. 10, pp. 2826-2834, Oct. 2005.
[21] K. A. Patwardhan and G. Sapiro, “Projection based image and video inpainting using wavelets,” in Proc. of the IEEE Int'l Conf. on Image Process., pp. 857-860, 2003.
[22] H. Yamauchi, J. Haber, and H. P. Seidel, “Image restoration using multiresolution texture synthesis and image inpainting,” in Proc. of the IEEE Int'l Conf. on Computer Graphics International, pp. 120-125, 2003.
[23] O. G. Guleryuz, “Nonlinear approximation based image recovery using adaptive sparse reconstructions and iterated denoising-part I: theory,” IEEE Trans. Image Processing, vol. 15, no. 3, pp. 539 - 554, Mar. 2006.
[24] O. G. Guleryuz, “Nonlinear approximation based image recovery using adaptive sparse reconstructions and iterated denoising-part II: adaptive algorithms,” IEEE Trans. Image Processing, vol. 15, no. 3, pp. 555-571, Mar. 2006.
[25] J. B. Kim, H. J. Kim, and S. Wachenfeld, “Restoration of regions occluded by a caption in TV scene,” in Proc. TENCON Conf., 2003.
[26] T. H. Tsai, C. L. Fang, and H. Y. Lin, “Progressive text regions inpainting based on edge detection and statistic method,” in Proc. Int'l Multi Conf. on Engineening and Computer Scientists, 2006.
[27] B. T. Chun, Y. Bae, and T. Y. Kim, “A method for original image recovery for caption areas in video,” in Proc. of the IEEE Int’l Conf. on Syst., Man, Cybernet., 1999.
[28] C. W. Lee, K. Jung, and H. J. Kim, “Automatic text detection and removal in video sequences,” Pattern Recognition Letters, vol. 24, no. 15, pp. 2607-2623, Nov. 2003.
[29] Y. Wexler, E. Shechtman, and M. Irani, “Space-time video completion,” IEEE Trans. Pattern Analysis and Machine Intelligence, vol. 29, no. 3, pp. 463-476, Mar. 2007.
[30] T. Shiratori, Y. Matsushita, S. B. Kang, and X. Tang, “Video completion by motion field transfer,” in Proc. of the IEEE Int'l Conf. on Computer Vision and Pattern Recognition, pp. 411-418, 2006.
[31] Y. Chen, Y. Hu, O. C. Au, H. Li, and C. W. Chen, “Video error concealment using spatio-temporal boundary matching and partial differential equation,” IEEE Trans. Multimedia, vol. 10, no. 1, pp. 2-15, Jan. 2008.
[32] K. A. Patwardhan, G. Sapiro, and M. Bertalmio, “Video inpainting under constrained camera motion,” IEEE Trans. Image Processing, vol. 16, no. 2, pp. 545-553, Feb. 2007.
[33] J. Jia, Y. W. Tai, T. P. Wu, and C. K. Tang, “Video repairing under variable illumination using cyclic motions,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 28, no. 5, pp. 832-83, May 2006.
[34] M. V. Venkatesh, S. C. S. Cheung, and J. Zhao, “Efficient object-based video inpainting,” Pattern Recognition Letters, vol. 30, no. 2, pp. 168-179, Jan. 2009.
[35] Y. Shen, F. Lu, X. Cao, and H. Foroosh, “Video completion for perspective camera under constrained motion,” in Proc. of the IEEE Int'l Conf. on Pattern Recognition, pp. 63-66, 2006.
[36] C. L. Fang and T. H. Tsai, “Advertisement video completion using hierarchical model,” in Proc. of the IEEE Int'l Conf. on Multimedia and Expo, 2008.
[37] J. Canny, “A computational approach to edge detection,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 8, no. 6, pp. 679-698, 1986.
[38] H. Li, D. Doermann, and O. Kia, “Automatic text detection and tracking in digital video,” IEEE Trans. Image Process., vol. 9, no. 1, pp. 147–156, Jan. 2000.
[39] A. K. Jain and B. Yu, “Automatic text location in images and video frames,” Pattern Recognit., vol. 31, no. 12, pp. 2055–2076, 1998.
[40] S. Antani, D. Crandall, and R. Kasturi, “Robust extraction of text in video,” in Proc. of the Int’l Conf. on Pattern Recognition., vol. 1, pp. 831–834, 2000.
[41] Y. Zhong, H.-J. Zhang, and A. K. Jain, “Automatic caption localization in compressed video,” in Proc. Int’l Conf. on Image Process., vol. 2, pp. 96–100, 1999.
[42] L. Agnihotri and N. Dimitrova, “Text detection for video analysis,” in Proc. IEEE Workshop Content-Based Access Image Video Libraries, pp. 109–113, 1999.
[43] T. Sato, T. Kanade, E. K. Hughes, and M. A. Smith, “Video OCR for digital news archive,” in Proc. IEEE Workshop Content-Based Access Image Video Database, 1998, pp. 52–60.
[44] V. Wu, R. Manmatha, and E. M. Riseman, “Textfinder: An automatic system to detect and recognize text in images,” IEEE Trans. on Pattern Analysis and Machine Intelligence, vol. 21, no. 11, pp. 1224–1229, Nov. 1999.
[45] M.R. Lyu, J. Song, and M. Cai, “A comprehensive method for multilingual video text detection, localization, and extraction”, IEEE Trans. Circuits Syst. Video Technol., vol. 15,no. 2, Feb. 2005, pp. 243 – 255.
[46] C. Wolf, J.-M. Jolion, and F. Chassaing, “Text localization, enhancement and binarization in multimedia documents” in Proc. of the IEEE Int'l Conf. on Pattern Recognition, 2002.
[47] D. Q. Zhang, B. L. Tseng, and S. F. Chang, “Accurate overlay text extraction for digital video analysis,” in Proc. of the Int’l Conf. Info. Tech. Research and Education, pp. 233–237, Aug. 2003.
[48] R. Lienhart and A. Wernicke, “Localizing and segmenting text in images, videos and web pages,” IEEE Trans. Circuits Syst. Video Technol., vol. 12, no. 4, pp. 256–268, Apr. 2002.
[49] X. Gao and X. Tang, “Automatic news video caption extraction and recognition,” in Proc. LNCS 1983: 2nd Int. Conf. Intell. Data Eng. Automated Learning Data Mining, Financial Eng., Intell. Agents, pp. 425–430, 2000.
[50] A. Wernicke and R. Lienhart, “On the segmentation of text in videos,” in Proc. of the IEEE Int’l Conf. on Multimedia and Expo, pp. 1511–1514, Jul. 2000.
[51] M. Cai, J. Song, and M. R. Lyu, “A new approach for video text detection,” in Proc. of the Int’l Conf. on Image Process., pp. 117–120, Sep. 2002.
[52] D. Chen, K. Shearer, and H. Bourlard, “Text enhancement with asymmetric filter for video OCR,” in Proc. of the Int’l Conf. Image Anal. Process. , pp. 192–197, 2001.
[53] B. T. Chun, Y. Bae, and T.-Y. Kim, “Text extraction in videos using topographical features of characters,” in Proc. of the IEEE Int’l Fuzzy Syst. Conf., vol. 2, pp. 1126–1130, 1999.
[54] S. Kwak, K. Chung, Y. Choi, “Video Caption Image Enhancement for an Efficient Character Recognition”, in Proc. of the Int’l Conf. on Pattern Recognition., vol. 2, pp. 2606–2609, 2000.
[55] F. Pescador, C. Sanz, M. J. Garrido, C. Santos, and R. Antoniello, “A DSP based IP set-top box for home entertainment,” IEEE Trans. on Consumer Electronics, vol. 52, no. 1, Feb. 2006.
[56] I. O. Kirenko, R. J. Vleuten, and L. Shao, “Optimizing scalable video compression for efficient implementation on a VLIW media processor,” IEEE Trans. on Multimedia, vol. 9, no. 2, Feb. 2007.
[57] B. Lei, W. Jin, J. Hu, and X. Zhang, “Embedded software optimization for AVS-P7 decoder real-time implementation on RISC core,” IEEE Trans. on Consumer Electronics, vol. 53, no. 3, Aug., 2007.
[58] L. K. Liu, Q. Liu, A. Natsey, K. A. Ross, J. R. Smith, and A. L. Varbanescu, “Digital media indexing on the cell processor,” in Proc. IEEE Int'l Conf. on Multimedia and Expo, 2007.
[59] G. Berger, R. Goedeken, and J. Richardson, “Motivation and implementation of a software H.264 real-time CIF encoder for mobile TV broadcast applications,” IEEE Trans. on Broadcasting, vol. 53, no. 2, Jun. 2007.
[60] W. Li, E. Li, N. Di, C. Dulong, T. Wang, and Y. Zhang, “On parallelization of a video mining system,” in Proc .of the IEEE Int'l Conf. on Multimedia and Expo, 2006.
[61] W. Li, X. Tong, and Y. Zhang, “Optimization and parallelization on a multimedia application,” in Proc. of the IEEE Int'l Conf. on Multimedia and Expo, 2007.
[62] X. Hu, U. Y. Ogras, N. H. Zamora, and R. Marculescu, “Data partitioning techniques for pervasive multimedia platforms,” in Proc. of the IEEE Int'l Conf. on Multimedia and Expo, 2004.
[63] M. Kang and W. Sung, “Memory access overhead reduction for a digital color copier implementation using a VLIW digital signal processor,” in Proc. of the IEEE Int'l Symposium on Circuits and Syst., 2005.
[64] P. Malani, Y. Tan, and Q. Qiu, “Resource-aware high performance scheduling for embedded MPSoCs with the application of MPEG decoding,” in Proc. of the IEEE Int'l Conf. on Multimedia and Expo, 2007.
[65] C. Steiger, H. Walder, and M. Platzner, “Operating systems for reconfigurable embedded platforms: online scheduling of real-time tasks,” IEEE Trans. on Computers, vol. 53, no. 11, Nov., 2004.
[66] S. P. Ierodiaconou, N. Dahnoun, and L. Q. Xu, “Implementation and optimization of a video object segmentation algorithm on an embedded DSP platform,” in Proc. Institution of Engineering and Technology Conf. on Crime and Security, 2006.
[67] “TMS320DM6446 digital media system-on-chip,” Datasheet (SPRS283E), Mar., 2007.
[68] J. Yang, D. Schonfeld, C. Chen, and M. Mohamed, “Online video stabilization based on particle filters,” in Proc. of the IEEE Int'l Conf. on Image Processing, 2006.
[69] M. Tico and M. Vehvilainen, “Constraint motion filtering for video stabilization,” in Proc. of the IEEE Int'l Conf. on Image Processing, 2005.
[70] Z. Juanjuan, G. Baolong, “Electronic image stabilization based on global feature tracking,” Journal of Systems Engineering and Electronics, vol. 19, no. 2, pp. 228-233, 2008.
[71] Chris Harris and Mike Stephens, “A combined corner and edge detector,” in Proc. of the Alvey vision conference, 1988.
[72] I. Tsubaki, T. Morita, T. Saito, and K. Aizawa, “An adaptive video stabilization method for reducing visually induced motion sickness,” in Proc. of the IEEE Int'l Conf. on Image Processing, 2005.
[73] H. C. Chang, S. H. Lai, and K. R. Lu, “A robust real-time video stabilization algorithm,” Journal of Visual Communication & Image Representation, vol. 17, no. 3, pp. 659-673, Jun. 2006.
[74] H. H. Chen, C. K. Liang, Y. C. Peng, and H. A. Chang, “Integration of digital stabilizer with video codec for digital video cameras,” IEEE Trans. Circuits Syst. Video Technol., vol. 17, no. 7, pp. 801-813, Jul. 2007.
[75] A. Bosco, A. Bruna, S. Battiato, G. Bella, and G. Pluglisi, “Digital video stabilization through curve warping techniques,” IEEE Trans. Consumer Electronics, vol. 54, no. 2, pp. 220-224, May 2008.
[76] A. U. Batur and B. Flinchbaugh, “Video stabilization with optimized motion estimation resolution,” Proc. of the IEEE Int'l Conf. on Image Processing, 2006.
[77] Y. Liu, M. Zou, Y. Cao, and X. Lu, “Dynamic Displacement field model used as a new camera motion model in video stabilization,” in Proc. of the IEEE Int'l Conf. on Consumer Electronics, 2007.
[78] Y. M. Liang, H. R. Tyan, S. L. Chang, H. Y. M. Liao, and S. W. Chen, “Video stabilization for a camcorder mounted on a moving vehicle,” IEEE Trans. Vehicular Technology, vol. 53, no. 6, pp. 1636-1648, Nov. 2004.
[79] J. A. Im, D. W. Kim, and K. S. Hong, “Digital video stabilization algorithm for cmos image sensor,” in Proc. of the IEEE Int'l Conf. on Image Processing, 2006.
[80] S. Auberger and C. Miro, “Digital Video Stabilization Architecture for Low Cost Devices,” in Proc. of the Int'l Symp. Conf. on Image and Signal Processing and Analysis, 2005.
[81] X. Xiaoshen, J. Hongxu, J. Liang, L. Dandan, and L. Bo, “A Multi-DSP System for High-Performance Video Applications,” in Proc. of IEEE Int'l Conf. on Communication Systems, 2008.
[82] N. Otsu, “A threshold selection method from gray-level histograms,” IEEE Trans. Syst., Man, Cybernet., vol. SMC-9, no. 1, pp. 62–66, Jan. 1979.
[83] M. Niskanen, O. Silven, and M. Tico, ”Video stabilization performance assessment,” in Proc. IEEE Int'l Conf. on Multimedia and Expo, 2006.
指導教授 蔡宗漢(Tsung-Han Tsai) 審核日期 2011-7-15
推文 facebook   plurk   twitter   funp   google   live   udn   HD   myshare   reddit   netvibes   friend   youpush   delicious   baidu   
網路書籤 Google bookmarks   del.icio.us   hemidemi   myshare   

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明