博碩士論文 975203012 詳細資訊




以作者查詢圖書館館藏 以作者查詢臺灣博碩士 以作者查詢全國書目 勘誤回報 、線上人數:35 、訪客IP:18.223.21.5
姓名 呂健益(Jian-Yi Lu)  查詢紙本館藏   畢業系所 通訊工程學系
論文名稱 以運動補償模型為基礎之移動式平台物件追蹤
(Object Tracking Using Motion Compensation Based Motion Model for Mobile Cameras)
相關論文
★ 應用於車內視訊之光線適應性視訊壓縮編碼器設計★ 以粒子濾波法為基礎之改良式頭部追蹤系統
★ 應用於空間與CGS可調性視訊編碼器之快速模式決策演算法★ 應用於人臉表情辨識之強健式主動外觀模型搜尋演算法
★ 結合Epipolar Geometry為基礎之視角間預測與快速畫面間預測方向決策之多視角視訊編碼★ 基於改良式可信度傳遞於同質區域之立體視覺匹配演算法
★ 以階層式Boosting演算法為基礎之棒球軌跡辨識★ 多視角視訊編碼之快速參考畫面方向決策
★ 以線上統計為基礎應用於CGS可調式編碼器之快速模式決策★ 適用於唇形辨識之改良式主動形狀模型匹配演算法
★ 基於匹配代價之非對稱式立體匹配遮蔽偵測★ 以動量為基礎之快速多視角視訊編碼模式決策
★ 應用於地點影像辨識之快速局部L-SVMs群體分類器★ 以高品質合成視角為導向之快速深度視訊編碼模式決策
★ 以運動補償模型為基礎之移動式相機多物件追蹤★ 基於匹配代價曲線特徵之遮蔽偵測之研究
檔案 [Endnote RIS 格式]    [Bibtex 格式]    [相關文章]   [文章引用]   [完整記錄]   [館藏目錄]   [檢視]  [下載]
  1. 本電子論文使用權限為同意立即開放。
  2. 已達開放權限電子全文僅授權使用者為學術研究之目的,進行個人非營利性質之檢索、閱讀、列印。
  3. 請遵守中華民國著作權法之相關規定,切勿任意重製、散佈、改作、轉貼、播送,以免觸法。

摘要(中) 電腦視覺在移動平台上的應用日趨發達,其中物件追蹤在手機等移動平台的應用中扮演重要角色。在移動平台上追蹤物件困難點在於,畫面中不僅有目標物件在移動,同時背景也不停變化,物件可能出現的範圍比靜止平台的物件追蹤要大上許多。當物件在前後畫面位置差距大時,對於物件移動的預測就顯得格外重要。因此,本論文提出適用在移動平台追蹤上的運動補償模型,透過粒子濾波器(PF),估測出補償相機移動量後之物件在二維影像中的運動量。粒子濾波器利用相機和物件的補償相機移動量後之物件在二維影像中的運動量,估測物件在二維影像上的運動量。相機的運動量,則使用SURF演算法計算。實驗結果顯示,我們提出的追蹤演算法應用在移動式平台上時,對於移動快速或不規則運動的物件皆能有良好的追蹤效果。
摘要(英) Robustness of visual tracking on mobile devices plays a key role in the success of emerging applications of robotics and cell phones. The prediction accuracy of object motion makes a great impact on the tracking accuracy. Therefore, this paper proposes to combine motion compensation and motion model to achieve fast and accurate estimation of object motion in 2-D images. At each time instant, the particle filter estimates the 2D object motion after motion compensation in the image based on the information of camera motionCamera motion is extracted with the aid of Speed-Up Robust Feature (SURF) algorithms. With our tracking algorithm, accurate tracking on mobile platforms can be achieved with a small number of particles. The experimental results show that our proposed tracking algorithm on mobile platforms performs well even if the rapid and irregular object motion exists.
關鍵字(中) ★ 運動補償模型
★ 移動平台
★ 物件追蹤
★ 粒子濾波器
關鍵字(英) ★ particle filter
★ motion model
★ mobile cameras
★ Visual tracking
論文目次 摘要 ii
Abstract iii
誌謝 iv
目錄 v
圖目錄 vii
表目錄 x
第一章 緒論 1
1.1 前言 1
1.2 研究動機 1
1.3 研究方法 3
1.4 論文架構 3
第二章 貝式估測器 3
2.1 貝氏濾波器(Bayesian Filter) 4
2.2 卡爾曼濾波器(Kalman Filter) 6
2.3 擴展式卡爾曼濾波器(Extended Kalman Filter) 8
2.4 粒子濾波器(Particle Filter) 9
2.5 以色彩為基礎的粒子濾波器運用在物件追蹤 11
2.5.1 運動模型(Motion Model) 11
2.5.2 量測模型(Measurement Model) 12
2.5.3 色彩模型更新與再取樣程序 14
2.6 總結 15
第三章 基於移動平台之物件追蹤技術現況 16
3.1 經由偵測的追蹤(Tracking-by-Detection) 16
3.2 運動補償為基礎的物件追蹤 19
3.3 總結 23
第四章 以運動補償為基礎的運動模型之物件追蹤系統 24
4.1 系統架構 24
4.2 以Speed-up Robust Feature為基礎計算全域運動特徵 28
4.3補償相機移動量後之物件在二維影像中的運動量的初始值 30
4.4 以運動補償模型估測補償相機移動量後之物件在二維影像中的運動量 32
4.5 運動模型的可調式高斯雜訊 34
4.6 結論 35
第五章 實驗結果與討論 36
5.1 實驗參數與測試影片 36
5.2.1 追蹤系統的準確度 37
5.2.2 追蹤系統的強健度 39
5.2.3 粒子權重差異度 40
5.2.4 系統計算複雜度 43
5.3 總結 43
第六章 結論與未來展望 44
參考文獻 45
參考文獻 [1] M. Isard and A. Blake, “Contour tracking by stochastic propagation of conditional density,” in Proceedings of European Conference on Computer Vision, April 1996.
[2] Y. Wu and T. S. Huang, “A co-inference approach to robust visual tracking,” in Proceedings of IEEE International Conference on Computer Vision, vol. 2, pp.26–33, July 2001.
[3] S. H. K. Zhou, R. Chellappa, and B. Moghaddam, “Visual tracking and recognition using appearance-adaptive models in particle filters,” IEEE Transactions on Image Processing, vol. 13, no. 11, pp. 1434-1456, November 2004.
[4] M. Isard and A. Blake, “ICONDENSATION: Unifying low-level and high-level tracking in a stochastic framework,” in Proceedings of European Conference on Computer Vision, vol. 1, pp. 767–781, 1998.
[5] N. Bouaynaya and D. Schonfeld, “On the optimality of motion-based particle filtering,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 19, no. 7, pp. 1068-1072, July 2009.
[6] K. Nummiaro, E. Koller-Meier, and L. V. Gool, “An adaptive color-based particle filter,”in Proceedings of International Conference on Image and Vision Computing, vol. 21, no. 1, pp. 99-110, January 2003.
[7] J. L. Yang, D. Schonfeld, and M. Mohamed, “Robust video stabilization based on particle filter tracking of projected camera motion,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 19, no. 7, July 2009.
[8] R. E. Kalman and R. S. Bucy, “New results in linear filtering and prediction theory,” ASME- Journal of Basic Engineering, vol. 83, pp. 95-107, 1961.
[9] N. Gordon, D. Salmond and A. Smith, “Novel approach to nonlinear/non-Gaussian Bayesian state estimation,” IEE Proceedings on Radar and Signal Processing, vol. 140, no. 2, pp. 107-113, 1993.
[10] G. Kitagawa, “Monte Carlo filter and smoother for non-Gaussian nonlinear state space models,” Journal of Computational and Graphical Statistics, vol. 5, no. 1, pp. 1-25, 1996.
[11] M. Isard and A. Blake, “Visual tracking by stochastic propagation of conditional density,” in Proceedings of European conference on Computer Vision, Cambridge, England, pp. 343–356, 1996.
[12] M. Isard and A. Blake, “CONDENSATION-conditional density propagation for visual tracking,” International Journal of Computer Vision, pp. 5–28, 1998.
[13] S. Thrun, W. Burgard, and D. Fox, Probabilistic Robotics, The MIT Press, Cambridge, Massachusetts, London, England, 2005.
[14] A. Lehuger, P. Lechat, and P. Perez, “An adaptive mixture color model for robust visual tracking, “in Proceedings of IEEE Computer Society Conference on Image Processing, pp. 573-576, 2006.
[15] K. Nummiaro, E. Koller-Meier, T. Svoboda, D. Roth, and L. Van Gool, “Color-based object tracking in multi-camera environments,” in Proceedings of DAGM Symposium on Pattern Recognition, pp. 591-599, 2003.
[16] P. Perez, C. Hue, J. Vermaak, and M. Gangnet, “Color-based probabilistic tracking,” in Proceedings of European Conference on Computer Vision, pp. 661-675, 2002.
[17] J. Ruiz-del-Solar and P. A. Vallejos, “Motion detection and tracking for an AIBO robot using camera motion compensation and Kalman filtering,” Lecture Notes in Computer Science - Springer, 2005.
[18] D. Stricker and T. Kettenbach, “Real-time and markerless vision-based tracking for outdoor augmented reality applications,” in Proceedings of IEEE and ACM International Symposium on Augmented Reality, 2001.
[19] A. Bulbul, Z. Cipiloglu, and T. Capin, “A face tracking algorithm for user interaction in mobile devices,” in Proceedings of International Conference on CyberWorlds, pp. 385-390, 2009.
[20] Y. Sugaya and K. Kanatani, “Extracting moving objects from a moving camera video sequence,” in Proceedings of 10th Symposium on Sensing via Imaging Information, pp. 279–284, 2004.
[21] J. Kang, I. Cohen, and G. Medioni, “Continuous multi-views tracking using tensor voting,” in Proceedings of the IEEE Workshop on Motion and Video Computing, pp. 181–186, Orlando, Florida, December 2002.
[22] R. Vidal, “Multi-subspace methods for motion segmentation from affine, perspective and central panoramic cameras,“in Proceedings of the 2005 IEEE International Conference on Robotics and Automation, pp. 1216–1221, 2005.
[23] M. D. Breitenstein, F. Reichlin, B. Leibe, E. Koller-Meier, and L. V. Gool,“ Markovian tracking-by-detection from a single, uncalibrated camera,” in Proceedings of IEEE CVPR PETS, June 2009.
[24] M. Andriluka, S. Roth, and B. Schiele, “People-tracking-by-detection and people-detection-by-tracking,” in Proceedings of IEEE CVPR, June 2008.
[25] J. Berclaz, F. Fleuret, and P. Fua, “ Robust people tracking with global trajectory optimization,” in Proceedings of IEEE CVPR, June 2006.
[26] S. Avidan, “Ensemble tracking,” in Proceedings of IEEE CVPR, vol. 2, pp. 494–501, June 2005.
[27] J. Shao, S.K. Zhou, and Q. Zheng, “Robust appearance-based tracking of moving object from moving platform,” Pattern Recognition, vol. 4, pp. 215 – 218, 2004.
[28] B. Jung and G. S. Sukhatme, “Real-time motion tracking from a mobile robot,” International Journal of Social Robotics, vol. 2, no. 1, pp. 63-78, March 2010.
[29] J. L. Yang, D. Schonfeld, and M. Mohamed, “Robust video stabilization based on particle filter tracking of projected camera motion,” IEEE Transactions on Circuits and Systems for Video Technology, vol. 19, no. 7, July 2009.
[30] J. Shi and C. Tomasi, “Good features to track,” in Proceedings of IEEE International Conference on Computer Vision and Pattern Recognition, pp. 593–600, June 1994.
[31] Y. Cai, N. de Freitas, and J.J. Little, “Robust visual tracking for multiple targets,” Lecture Notes in Computer Science - Springer, 2006.
[32] A. Yilmaz, K. Shafique, N. Lobo, X. Lin, T. Olson, and M. Shah, “Target-tracking in FLIR imagery using mean-shift and global motion compensation,“ in Workshop on Computer Vision Beyond the Visible Spectrum, pp. 54-58, 2001.
[33] R. Collins, X. Zhou, and S. K. Teh, “An open source tracking testbed and evaluation web site,” in Proceedings of IEEE International Workshop on Performance Evaluation of Tracking and Surveillance, January 2005.
[34] D.G. Lowe, “Object recognition from local scale-invariant features,” in IEEE Proc. International Conference on Computer Vision, pp. 1150–1157, September 1999.
[35] H. Bay, T. Tuytelaars, and L. V. Gool, “SURF: speeded up robust features,” in Proc. European Conference on Computer Vision, Graz, vol. 13, pp. 404–417, 2006.
[36] M. Kristan, “A local-motion-based probabilistic model for visual tracking,” Pattern Recognition, vol. 42, no. 9, pp. 2160-2168, January 2009.
[37] D.W. Scott, Multivariate density estimation, Wiley, New York, 1992.
[38] S. Birchfield, “Elliptical head tracking using intensity gradients and color histograms,” in Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Santa Barbara, California, pages 232-237, June 1998.
[39]http://groups.inf.ed.ac.uk/vision/CAVIAR/CAVIARDATA1/
指導教授 唐之瑋(Chih-Wei Tang) 審核日期 2010-7-27
推文 facebook   plurk   twitter   funp   google   live   udn   HD   myshare   reddit   netvibes   friend   youpush   delicious   baidu   
網路書籤 Google bookmarks   del.icio.us   hemidemi   myshare   

若有論文相關問題,請聯絡國立中央大學圖書館推廣服務組 TEL:(03)422-7151轉57407,或E-mail聯絡  - 隱私權政策聲明