用户名: 密码: 验证码:
基于Web图像的视觉模式挖掘研究
详细信息    本馆镜像全文|  推荐本文 |  |   获取CNKI官网全文
摘要
随着网络技术的飞速发展,Web上拥有的图像资源已经越来越丰富。这个巨大的图像数据库中蕴藏着大量对用户有价值的信息。图像挖掘技术致力于对海量图像数据自动分析处理,以此获取有意义的模式和知识。基于Web的图像内容挖掘是近年来多媒体数据挖掘领域的热点研究方向之一。本文围绕Web图像的视觉模式发现和提取展开,重点研究了基于显著性和基于语义的视觉模式挖掘方法,并将这两种方法用于图像检索中的重排序。
     首先,本文通过研究视觉选择性注意机制的计算模型,提出了一种基于统计学习的多尺度视觉显著性模型建立方法。在此基础上,将该模型用于Web图像挖掘,提出了一种基于显著性的视觉模式提取方法。文中给出了显著图和非显著图的定义,把通常考虑显著性问题的视角从区域扩展到整幅图像。同时,进行了多尺度相关性的讨论,使用多尺度的表示方法能对图像进行更为精确地描述。根据构建的显著图和非显著图像数据库,分析了这两类图像所呈现的不同视觉特性,以此选取颜色、边缘、纹理和图像要旨(Gist)四个不同的底层特征来训练视觉显著性模型。实验部分,对模型进行了客观性的定量分析和主观性的眼动实验,验证了该模型理论假设的正确性及在实际应用中对显著图检测的有效性。
     其次,本文研究基于语义的视觉模式挖掘问题,提出了一种无监督的方法对来自于Web的图像自动进行聚类分析,从而提取出特定语义概念的主要视觉模式。该算法主要针对物体类概念,对于给定的查询关键词,充分利用了丰富的Web图像资源,无需人工干预自动地挖掘所包含的主要视觉模式。文中给出了具有视觉一致性图像的定义:当使用Web图像搜索引擎时,在返回结果的前几页中出现频率较高同时视觉上相似的图像往往与用户的查询主题相关。利用这些图像的一致性信息,挖掘特定语义概念的主要视觉模式。此外,Google图像搜索引擎提供了剪贴画检索功能。剪贴画不仅具有干净的背景,而且最大程度上反映了物体的基本形状。利用剪贴画的特有属性,可以方便有效地提取出有价值的图像底层特征信息。基于以上两点,本文提出一种基于语义的视觉模式挖掘的新方法。实验结果表明,该方法充分利用了图像集合中的一致性信息和剪贴画的特有属性,能有效挖掘出特定语义概念的主要视觉模式。利用该方法挖掘出的视觉模式,不仅能用于提升图像聚类、浏览和检索的性能,同时也能应用于物体分类、检测、识别等领域。
     最后,本文研究搜索引擎返回的原始图像的重排序问题,提出了一种基于视觉显著性和一致性的图像重排序算法。在Web图像搜索应用中,视觉一致性的图像在重排序时应给予更高的相关性分值。此外,从视觉角度出发,视觉显著的图像更能吸引人的注意。同时也观察到在搜索引擎返回结果的前几页中,视觉显著的图像更有可能与用户查询相关。从以上两点出发,本文提出一种新的基于随机游走的融合方法,将视觉显著性和一致性结合起来用于图像的重排序。实验结果表明,该算法能有效地提升搜索引擎的检索性能,将视觉上显著的且与查询主题密切相关的图像优先返回给用户。
With the rapid development of network technology, web image resourceshave become a huge image database which provides users with abundant valu-able information. In order to obtain meaningful patterns and knowledge, theimage mining technology has been applied in the analyzing and processing oflarge scale image data. In the field of multimedia data mining, web-based im-age content mining has been one of the hottest research topics in resent years.The present study dues with web image visual pattern discovery and extraction.It mainly investigates the methods of saliency-based and concept-based visualpattern mining, and uses the methods to re-rank images in image retrieval.Firstly, the dissertation probes into the computational model of visual se-lective attention mechanism, and provide a method to build multiscale visualsaliency model based on statistical learning. By using such a model, the miningmethod of saliency-based visual pattern is proposed in the study. The authordefines salient image and cluttered image based on the saliency of a whole imagerather then a region. And, multiscale relevance is analyzed in order to give moreaccurate descriptions to images. Based on the databases of salient images andcluttered images, the author analyzes the perceptual di?erences of the two classesof images, and chooses color, edge, texture, Gist as the four features to train thevisual saliency model. In the experiments, quantitative analysis is conducted tothe model. Eye movement test is done to test the hypothesis about the modeltheory and the applicability of the model in saliency image detection.
     Secondly, the dissertation discusses the problem of concept-based visual pat-tern mining. A new method of automatic clustering analysis for web images isproposed to extract the main visual pattern according to specific concept. Devel- oped for the concept of object category and by making full use of rich web imageresources, the algorithm is able to mine the main visual pattern of images withouthuman intervention for a given query. The author defines the visual consistentimage: When using a web image search engine,the images closely related to thequery are often visually similar and occur most frequently in the first few webpages. Based on such a fact, the proposed method mines the main visual patternby using the visual consistency. Besides, Google image search engine providesclip art retrieval function. The clip arts often have clear backgrounds, and re?ectthe basic shapes of objects to the greatest extent. Such special attributes arehelpful for the extraction of valuable low-level features of objects. Grounded onthe two points, the dissertation proposes a new visual pattern mining methodbased on semantic concept. The experiment shows that the proposed methodcan be able to find the main visual pattern of specific concept e?ciently by uti-lizing the consistency of image and the special attributes of clip arts. This visualpattern can not only be used to improve the performance of image clustering,browsing and retrieval, but also be applied in some other fields such as objectclassification, detection and recognition.
     Lastly, we propose a new algorithm for image re-ranking in web image searchapplications. The proposed method focuses on investigating the following twomechanisms: visual consistency and visual saliency. In web image search cases,when re-ranking images, these visually consistent images would be given higherranks. Besides, from visual aspect, it is obvious that salient images would beeasier to catch users’eyes and it is observed that these visually salient images inthe front pages are often relevant to the user’s query. By integrating the above twomechanisms, our method can e?ciently re-rank the images from search enginesand obtain a more satisfactory search result. Experimental results on a real-world web image dataset demonstrate that our approach can e?ectively improvethe performance of image retrieval. These images which are salient and closelyrelated to the query would be given priority to return for users.
引文
[1] J. Han and M. Kamber,《数据挖掘:概念与技术》,机械工业出版社,北京, 2007.
    [2] M.C. Burl,“Mining NASA’s large image collections,”in Joint StatisticalMeetings, Session on Mining Large Datasets, 2000.
    [3] J. Zhang, W. Hsu, and M. Lee,“Image mining: Issues, frameworks andtechniques,”in Proceedings of the Second International Workshop on Mul-timedia Data Mining, 2001.
    [4] W. Hsu, M. Lee, and J. Zhang,“Image mining: Trends and developments,”Journal of Intelligent Information Systems, vol. 19, no. 1, 2002.
    [5]陈久军,基于统计学习的图像语义挖掘研究, Ph.D. thesis,浙江大学,杭州, 2006.
    [6]段曼妮,图像挖掘在图像检索中的应用, Ph.D. thesis,中国科学技术大学,合肥, 2009.
    [7] U. Fayyad and P. Smyth,“Image database exploration: Progress and chal-lenges,”in Proceedings of knowledge discovery in databases workshop, 1993,pp. 14–27.
    [8] M. C. Burl, C. Fowlkes, J. Roden, A. Stechert, and S. Mukhtar,“Diamondeye: A distributed architecture for image data mining,”in SPIE AeroSenseConference on Data Mining and Knowledge Discover, 1999.
    [9] J. Roden, M. C. Burl, and C. Fowlkes,“The diamond eye image miningsystem,”in Proceedings of the 11th International Conference on Scientificand Statistical Database Management, 1999.
    [10] M. C. Burl, C. Fowlkes, and J.Roden,“Mining for image content,”inSCI-ISAS, 1999.
    [11] A. Kitamoto,“Data mining for typhoon image collection,”in Proceedingsof the Second International Workshop on Multimedia Data Mining, 2001,pp. 68–77.
    [12] A. Kitamoto,“Spatio-temporal data mining for typhoon image collection,”Journal of Intelligent Information Systems, vol. 19, no. 1, 2002.
    [13] R. Guadagnina, L. Santanab, E. Fernedaa, and H. Pradoa,“Improvingimage mining through geoprocessing,”Journal of Intelligent InformationSystems, vol. 20, no. 1, pp. 81–85, 2010.
    [14] A. Kusiak, K. H. Kernstine, J. A. Kern, K. A. McLaughlin, and T. L. Tseng,“Data mining: Medical and engineering case studies,”in Proceedings of theIIE Research 2000 Conference, 2000, pp. 1–7.
    [15] M. Kakimoto, C. Morita, and H. Tsukimoto,“Data mining from functionalbrain images,”in Proceedings of the International Workshop on MultimediaData Mining, 2000.
    [16] M. Antonie, O. R. Zayane, and A. Coman,“Application of data miningtechniques for medical image classification,”in Proceedings of the Interna-tional Workshop on Multimedia Data Mining, 2001.
    [17] A. K. Mohanty and S. K. Lenka,“E?cient image mining technique forclassification of mammograms to detect breast cancer,”Special Issue ofIJCCT, vol. 2, 2010.
    [18] P. Rajendran and M.Madheswaran,“Hybrid medical image classificationusing association rule mining with decision tree algorithm,”Journal OfComputing, vol. 2, no. 1, 2010.
    [19] P. Rajendran and M.Madheswaran,“Novel fuzzy association rule imagemining algorithm for medical decision support system,”International Jour-nal of Computer Applications, vol. 1, no. 20, 2010.
    [20] K. Yanai,“Web image mining: can we gather visual knowledge for imagerecognition from the web?,”in Proceedings of ICICS-PCM, 2003, vol. 1,pp. 186–190.
    [21] K. Yanai,“Web image mining toward generic image recognition,”in Pro-ceedings of the international conference on World Wide Web, 2003.
    [22] K. Yanai and K. Barnard,“Probabilistic web image gathering,”in Proceed-ings of ACM SIGMM International Workshop on Multimedia InformationRetrieval, 2005, pp. 57–64.
    [23] K. Yanai,“Automatic web image selection with a probabilistic latent topicmodel,”in Proceedings of the international conference on World Wide Web,2008, pp. 1237–1238.
    [24] K. Yanai and K. Barnard,“Finding visual concepts by web image mining,”in Proceedings of the international conference on World Wide Web, 2006,pp. 923–924.
    [25] Z. Chen, W. Liu, F. Zhang, M. Li, and H. J. Zhang,“Web mining for webimage retrieval,”Journal of the American Society for Information Scienceand Technology, vol. 52, no. 10, 2001.
    [26] C. Zhang, J. Liu, H. Lu, and S. Ma,“Web image mining using conceptsensitive markov stationary features,”in Proceedings of ICME, 2009, pp.462–465.
    [27] W. Premchaiswadi and A. Tungkatsathan,“On-line content-based im-age retrieval system using joint querying and relevance feedback scheme,”WSEAS Transactions on Computers, vol. 9, no. 5, 2010.
    [28] B. Ni, Z. Song, and S. Yan,“Web image mining towards universal ageestimator,”in Proceedings of the seventeen ACM international conferenceon Multimedia, 2009, pp. 85–94.
    [29] M. Y. Eldib and H. M. Onsi,“Web image mining age estimation frame-work,”ICGST International Journal on Graphics, Vision and Image Pro-cessing, vol. 11, no. 1, pp. 1–8, 2011.
    [30] Y. Liu, T. Mei, and X. S. Hua,“Crowdreranking: exploring multiple searchengines for visual search reranking,”in Proceedings of ACM Special InterestGroup on Information Retrieval, 2009, pp. 500–507.
    [31]刘媛,视频搜索结果的重排序研究, Ph.D. thesis,中国科学技术大学,合肥, 2009.
    [32] H. Zitouni, S. Sevil, D. Ozkan, and P. Duygulu,“Re-ranking of web imagesearch results using a graph algorithm,”in Proceedings of InternationalConference on Pattern Recognition, 2008, pp. 1–4.
    [33] Y. Jing and S. Baluja,“Visualrank: Applying pagerank to large-scale imagesearch,”IEEE Transactions on Pattern Analysis and Machine Intelligence,vol. 30, pp. 1877–1890, 2008.
    [34] X. Tian, L. Yang, J. Wang, Y. Yang, X. Wu, and X. S. Hua,“Bayesian videosearch reranking,”in Proceedings of ACM Multimedia, 2008, pp. 131–140.
    [35] A. Natsev, M. R. Naphade, and J. Tesic,“Learning the semantics of mul-timedia queries and concepts from a small number of examples,”in Pro-ceedings of ACM Multimedia, 2005, pp. 598–607.
    [36] R. Yan, A. Hauptmann, and R. Jin,“Multimedia search with pseudo-relevance feedback,”in Proceedings of CIVR Workshop, 2003.
    [37] N. Ben-Haim, B. Babenko, and S. Belongie,“Improving web-based imagesearch via content based clustering,”in Proceedings of CVPR Workshop,SLAM, 2006.
    [38] Winston H. Hsu, Lyndon S. Kennedy, and Shih-Fu Chang,“Video searchreranking via information bottleneck principle,”in Proceedings of ACM.Multimedia.
    [39] Winston H. Hsu, Lyndon S. Kennedy, and Shih-Fu Chang,“Rerankingmethods for visual search,”IEEE MultiMedia, vol. 14, pp. 14–22, 2007.
    [40] J. G. Carbonell et al,“Translingual information retrieval: A comparativeevaluation,”in International Joint Conference on Artificial Intelligence,1997.
    [41] A. Amir et al,“IBM research TRECVID-2005 video retrieval system,”inIn TRECVID Workshop, 2005.
    [42] R. Fergus, L .Fei-Fei, P. Perona, and A. Zisserman,“Learning object cat-egories from google’s images search,”in Proceedings of the IEEE Interna-tional Conference on Computer Vision (ICCV), 2005.
    [43] O. R. Za¨?ane, J. Han, Z. Li, and J. Hou,“Mining multimedia data,”inProceedings of the conference of the Centre for Advanced Studies on Col-laborative, 1998.
    [44] J. Zhang, W. Hsu, and M. L. Lee,“An information-driven framework forimage mining,”in Proceedings of the 12th International Conference onDatabase and Expert Systems Applications, 2001.
    [45] R. Jain, R. Kasturi, and B. G. Schunck, Machine Version, MIT Press,1995.
    [46] J. Barros, J. French, W. Martin, P. Kelly, and J. M. White,“Indexing mul-tispectral images for content-based retrieval,”in Proc. SPIE: 23rd AIPRWorkshop: Image and Information Systems, 1994, vol. 2368, pp. 25–36.
    [47]吴楠,宋方敏,“一种基于图像高层语义信息的图像检索方法,”中国图象图形学报, vol. 11, no. 12, 2006.
    [48] M. Stricker and M. Orengo,“Similarity of color images,”in Proceedings ofSPIE Storage and Retrieval for Image and Video Databases, 1995.
    [49] J. Huang, S. R. Kumar, M. Mitra, W. Zhu, and R. Zabih,“Image indexingusing color correlograms,”in Proceedings of the Conference on ComputerVision and Pattern Recognition, 1997, pp. 762–768.
    [50] H. Freeman,“Shape description via the use of critical points,”PatternRecognition, vol. 10, no. 3, pp. 159–166, 1978.
    [51] M. A. Fischler and H. C. Wolf,“Locating perceptually salient points onplanar curves,”IEEE Transactions on Pattern Analysis and Machine In-telligence, vol. 16, no. 2, pp. 113–129, 1994.
    [52] F. S. Cohen, Z. Huang, and Z. Yang,“Invariant matching and identificationof curves using b-splines curve representation,”IEEE Transactions onImage Processing, vol. 4, no. 1, pp. 1–10, 1995.
    [53] P. Saint-Marc, H. Rom, and G. Medioni,“B-spline contour representationand symmetry detection,”IEEE Transactions on Pattern Analysis andMachine Intelligence, vol. 15, no. 11, pp. 1191–1197, 1993.
    [54] C. T. Zahn and R. Z. Roskies,“Fourier descriptors for plane closed curves,”IEEE Transactions on Computers, vol. c-21, no. 3, pp. 269–281, 1972.
    [55] E. Persoon and K. S. Fu,“Shape discrimination using fourier descriptors,”IEEE Transactions on System, Man, Cybernetics, vol. 7, no. 3, pp. 170–179, 1977.
    [56] H. Kauppinen, T. Sepp¨anen, and M. Pietik¨ainen,“An experimental com-parison of autoregressive and fourier-based descriptors in 2d shape classifi-cation,”IEEE Transactions on Pattern Analysis and Machine Intelligence,vol. 17, no. 2, pp. 201–207, 1995.
    [57] H. Yang, S. Lee, and K. Lee,“Recognition of 2d object contours usingstarting-point-independent wavelet coe?cient matching,”Journal of VisualCommunication and Image Representation, vol. 9, no. 2, pp. 171–181, 1998.
    [49] J. Huang, S. R. Kumar, M. Mitra, W. Zhu, and R. Zabih,“Image indexingusing color correlograms,”in Proceedings of the Conference on ComputerVision and Pattern Recognition, 1997, pp. 762–768.
    [50] H. Freeman,“Shape description via the use of critical points,”PatternRecognition, vol. 10, no. 3, pp. 159–166, 1978.
    [51] M. A. Fischler and H. C. Wolf,“Locating perceptually salient points onplanar curves,”IEEE Transactions on Pattern Analysis and Machine In-telligence, vol. 16, no. 2, pp. 113–129, 1994.
    [52] F. S. Cohen, Z. Huang, and Z. Yang,“Invariant matching and identificationof curves using b-splines curve representation,”IEEE Transactions onImage Processing, vol. 4, no. 1, pp. 1–10, 1995.
    [53] P. Saint-Marc, H. Rom, and G. Medioni,“B-spline contour representationand symmetry detection,”IEEE Transactions on Pattern Analysis andMachine Intelligence, vol. 15, no. 11, pp. 1191–1197, 1993.
    [54] C. T. Zahn and R. Z. Roskies,“Fourier descriptors for plane closed curves,”IEEE Transactions on Computers, vol. c-21, no. 3, pp. 269–281, 1972.
    [55] E. Persoon and K. S. Fu,“Shape discrimination using fourier descriptors,”IEEE Transactions on System, Man, Cybernetics, vol. 7, no. 3, pp. 170–179, 1977.
    [56] H. Kauppinen, T. Sepp¨anen, and M. Pietik¨ainen,“An experimental com-parison of autoregressive and fourier-based descriptors in 2d shape classifi-cation,”IEEE Transactions on Pattern Analysis and Machine Intelligence,vol. 17, no. 2, pp. 201–207, 1995.
    [57] H. Yang, S. Lee, and K. Lee,“Recognition of 2d object contours usingstarting-point-independent wavelet coe?cient matching,”Journal of VisualCommunication and Image Representation, vol. 9, no. 2, pp. 171–181, 1998.
    [68] M. K. Mandal, T. Aboulnasr, and S. Panchanathan,“Fast wavelet his-togram techniques for image indexing,”Journal of Computer Vision andImage Understanding, vol. 75, no. 1-2, pp. 99–110, 1999.
    [69] T. Ojala, M. Pietik¨ainen, and T. M¨aenpa¨¨a,“Multiresolution gray scaleand rotation invariant texture analysis with local binary patterns,”IEEETransactions on Pattern Analysis and Machine Intelligence, vol. 24, pp.971–987, 2002.
    [70] J. Wu and J. M. Rehg,“Where am I : Place instance and category recog-nition using spatial pact,”in Proceedings of the IEEE Computer SocietyConference on Computer Vision and Pattern Recognition (CVPR), 2008.
    [71] S. K. Chang,“Iconic indexing by 2-d strings,”IEEE Transactions onPattern Analysis and Machine Intelligence, vol. 9, no. 3, pp. 413–428, 1987.
    [72] S. Lazebnik, C. Schmid, and J. Ponce,“Beyond bags of features: Spatialpyramid matching for recognizing natural scene categories,”in Proceedingsof the IEEE Computer Society Conference on Computer Vision and PatternRecognition (CVPR), 2006, pp. 2169–2178.
    [73] A. Vailaya, M. Figueiredo, A. Jain, and H. Zhang,“Image classification forcontent-based indexing,”IEEE Transactions on Image Processing, vol. 10,no. 1, pp. 117–130, 2001.
    [74] T. Hofmann,“Unsupervised learning by probabilistic latent semantic anal-ysis,”Machine Learning, vol. 42, pp. 177–196, 2001.
    [75] J. Sivic, B. C. Russell, A. A. Efros, A. Zisserman, and W. T. Freeman,“Discovering objects and their location in images,”in Proceedings of In-ternational Conference on Computer Vision, 2005, vol. 1, pp. 370–377.
    [76] D. Blei, Y. Andrew, and M. Jordan,“Latent dirichlet allocation,”Journalof Machine Learning Research, vol. 3, pp. 993–1020, 2003.
    [77] L. Fei-Fei and P. Perona,“A bayesian hierarchical model for learning nat-ural scene categories,”in Proceedings of the fifth Berkeley Symposium onMathematical Statistics and Probability, 2005, pp. 524–531.
    [78] C. Cortes and V. Vapnik,“Support-vector networks,”Machine Learning,vol. 20, no. 3, pp. 273–297, 1995.
    [79] K. Goh, E. Chang, and B. Li,“Using on-class and two-class svms formulticlass image annotation,”IEEE Transactions on Knowledge and DataEngineering, vol. 17, no. 10, pp. 1333–1346, 2005.
    [80] W. Zhang, B. Yu, G. Zelinsky, and D. Samaras,“Object class recognitionusing multiple layer boosting with heterogeneous features,”in IEEE Con-ference on Computer Vision and Pattern Recognition, 2005, pp. 323–330.
    [81] A. Opelt, A. Pinz, M. Fussenegger, and P. Auer,“Generic object recogni-tion with boosting,”IEEE Transactions on Pattern Analysis and MachineIntelligence, vol. 28, no. 3, pp. 416–431, 2008.
    [82] J.La?erty, A.McCallum, and F.Pereira,“Conditional random fields: Prob-abilistic models for segmenting and labeling sequence data,”in Proceedingsof 18th International Conference on Machine Learning, 2001, pp. 282–289.
    [83] J. B. MacQueen,“Some methods for classification and analysis of multi-variate observations,”in Proceedings of the fifth Berkeley Symposium onMathematical Statistics and Probability, 1967, vol. 1, pp. 281–297.
    [84] Q. Zhang and I. Couloigner,“A new and e?cient k-medoid algorithmfor spatial clustering,”in Computational Science and Its Applications–ICCSA, 2005, pp. 181–189.
    [85] F. Jurie and B. Triggs,“Creating e?cient codebooks for visual recogni-tion,”in Proceedings of IEEE International Conference on Computer Vi-sion, 2005, vol. 1, pp. 604–610.
    [86] T. Zhang and R. Ramakrishnan,“Birch: An e?cient data clusteringmethod for very large databases,”in Proceedings of ACM SIGMOD Con-ference, 1996, pp. 103–114.
    [87] S. Guha, R. Rastogi, and K. Shim,“Cure: An e?cient clustering algorithmfor large database,”in Proceedings of ACM SIGMOD Conference, 1998,pp. 73–84.
    [88] S. Guha, R. Rastogi, and K. Shim,“Rock: A robust clustering algorithmfor categorical attributes,”in Proceeding of International Conference DataEngineering, 1999, pp. 512–521.
    [89] M. Ester, H. Kriegel, J. Sander, and X. Xu,“A density-based algorithm fordiscovering clusters in large spatial databases with noise,”in Proceedingsof international conference on knowledge discovery and data mining, 1996,pp. 226–231.
    [90] R. Sanguthevar,“E?cient parallel hierarchical clustering algorithms,”IEEE Transactions on Parallel and Distributed Systems, vol. 16, no. 6,pp. 497–502, 2005.
    [91]薛向阳,“基于内容的多媒体和跨媒体信息检索技术,”世界科学, vol. 12,2005.
    [92]高永英,章毓晋,罗云,“基于目标语义特征的图像检索系统,”电子与信息学报, vol. 25, no. 10, 2003.
    [93] R. Desimone and J. Duncan,“Neural mechanisms of selective visual atten-tion,”Annual Review of Neuroscience, vol. 18, pp. 193–222, 1995.
    [94]刘伟,张宏,童勤业,“视觉注意计算模型及其在自然图像压缩中的应用,”浙江大学学报(工学版), vol. 41, no. 4, 2007.
    [95] S. W. Ban, M. Lee, and H. S. Yang,“A face detection using biologically mo-tivated bottom-up saliency map model and top-down perception model,”Neurcomputing, vol. 56, pp. 475–480, 2004.
    [96] A. L. Rothenstein and J. K. Tsotsos,“Attention links sensing to recogni-tion,”Image Vision and Computing, vol. 26, pp. 114–126, 2008.
    [97] U. Rutishauser, D. Walther, C. Koch, and P. Perona,“Is bottom-up at-tention useful for object recognition?,”in Proceeding of the 2004 IEEEComputer Society Conference on Computer Vision and Pattern Recogni-tion, 2004.
    [98] J. J. Bonaiuto and L. Itti,“Combining attention and recognition for rapidscene analysis,”in Proceeding of the 2004 IEEE Computer Society Confer-ence on Computer Vision and Pattern Recognition, 2005.
    [99] Y. Sun and R. Fisher,“Object-based visual attention for computer vision,”Artificial Intelligence, vol. 146, pp. 77–123, 2003.
    [100] A. P. Bradley and W. M. Stentiford,“Visual attention for region of in-terest coding in jpeg 2000,”Journal of Visual Communication & ImageRepresentation, pp. 232–250, 2003.
    [101] R. Datta, J. Li, and J. Z. Wang,“Content-based image retrieval approachesand trends of the new age,”in Proceedings of International Workshop onMultimedia Information Retrieval, 2005, pp. 253–262.
    [102] O. Marques, L. M. Mayron, G. B. Borba, and H. R. Gamba,“An attention-driven model for grouping similar images with image retrieval applications,”Journal on Advances in Signal Processing, 2007.
    [103] A. Bamidele, F. W. M. Stentiford, and J. Morphett,“An attention-basedapproach to content-based image retrieval,”BT Technology Journal, vol.22, pp. 151–160, 2004.
    [104] H. Fu, Z. Chi, and D. Feng,“Attention-driven image interpretation withapplication to image retrieval,”Pattern Recognition, vol. 39, pp. 1604–1621,2006.
    [105] A. Faro, D. Giordano, C. Pino, and C. Spampinato,“Visual attention forimplicit relevance feedback in a content based image retrieval,”in Proceed-ings of the 2010 Symposium Eye-Tracking Research & Applications, 2010,pp. 73–76.
    [106] O. K. Oyekoya and F. W. M. Stentiford,“A new interface for visual explo-ration,”BT Technology Journal, vol. 24, no. 3, pp. 57–66, 2006.
    [107] O. K. Oyekoya and F. W. M. Stentiford,“Perceptual image retrieval usingeye movements,”International Journal of Computer Mathematics, vol. 84,no. 9, pp. 1379–1391, 2007.
    [108] C. Koch and S. Ulman,“Shifts in selective visual attention: toward theunderlying neural circuitry,”Hum Neurobiol, vol. 4, no. 4, pp. 219–227,1985.
    [109] R. Milanese, Detecting salient regions in an image: From biological ev-idence to computer implementation, Ph.D. thesis, University of Geneva,Switzerland, 1993.
    [110] L. Itti, C. Koch, and E. Niebur,“A model of saliency-based visual attentionfor rapid scene analysis,”IEEE Trans. on Pattern Analysis and MachineIntelligence, vol. 20, no. 11, pp. 1254–1259, 1998.
    [111] P. R. N. Rao and D. H. Ballard,“An active vision architecture based onicon representations,”Artificial Intelligence Journal, vol. 78, 1995.
    [112]张鹏,图像信息处理中的选择性注意机制研究, Ph.D. thesis,国防科学技术大学,长沙, 2004.
    [113] L. Itti and C. Koch,“A saliency-based search mechanism for overt andcovert shifts of visual attention,”IEEE Transactions on Pattern Analysisand Machine Intelligence, vol. 40, pp. 1489–1506, 2000.
    [114] A. Dimitri, Chernyak, and L. Strark,“Top-down guided eye movements,”IEEE Transactions on Systems, Man and Cybernetics, vol. 31, 2001.
    [115] V. Navalpakkam and L. Itti,“Modeling the in?uence of task on attention,”Vision Researchs, vol. 45, no. 2, pp. 205–231, 2005.
    [116] S. Frintrop, G. Backer, and E. Rome,“Selecting what is important: Train-ing visual attention,”in Proceedings of KI, 2005, pp. 351–365.
    [117] S. Frintrop, G. Backer, and E. Rome,“Goal-directed search with a top-down modulated computational attention system,”Pattern Recognition,vol. 3663, pp. 117–124, 2005.
    [118] S. Frintrop and A. B. Cremers,“Top-down attention supports visual loopclosing,”in Proceedings of ECMR, 2007.
    [119] S. Frintrop, P. Jensfelt, and H. Christensen,“Pay attention when selectingfeatures,”in Proceedings of ICPR, 2006.
    [120] S. Frintrop, P. Jensfelt, and H. Christensen,“Attentional robot localizationand mapping,”in ICVS Workshop WCAA, 2007.
    [121] B. Rasolzadeh and J. O. Eklundh,“An attentional system combining top-down and bottom-up in?uences,”in International Workshop on Attentionin Cognitive Systems, 2007.
    [122] I. Biederman, R. J. Mezzanotte, and J. C. Rabinowitz,“Scene perception:detecting and judging objects undergoing relational violations,”Cognitivepsychology, vol. 14, no. 2, pp. 143–177, 1982.
    [123] S. E. Palmer,“The e?ects of contextual scenes on the identification ofobjects,”Memory & Cognition, vol. 3, no. 5, pp. 519–526, 1975.
    [124] A. Torralba,“Modeling global scene factors in attention,”Journal of theOptical Society of America A: Optics and Image Science, and Vision, vol.20, no. 7, pp. 1407–1418, 2003.
    [125] J. M. Henderson and A. Hollingworth,“High-level scene perception,”An-nual Review of Psychology, vol. 50, pp. 243–271, 1999.
    [126] R. A. Rensink, J. K. O’Regan, and J. J. Clark,“To see or not to see: Theneed for attention to perceive changes in scenes,”Psychological Science,vol. 8, no. 5, pp. 368–373, 1997.
    [127] R. A. Rensink,“The dynamic representation of scenes,”Visual Cognition,vol. 7, pp. 17–42, 2000.
    [128] P. G. Schyns and A. Oliva,“From blobs to boundary edges: evidence fortime and spatial scale dependent scene recognition,”Psychol. Sci., vol. 5,pp. 195–200, 1994.
    [129] H. Arsenio, A. Oliva, and J. M. Wolfe,“Exorcising ghosts in repeated visualsearch,”J. Vision, vol. 2, 2002.
    [130] P. J. Burt,“Fast filter transforms for image processing,”Computer Vision,Graphics and Image Processing, 1981.
    [131] A.K. Jain and A. Vailaya,“Image retrieval using color and shape,”PatternRecognition, vol. 29, pp. 1233–1244, 1996.
    [132] C. Siagian and L. Itti,“Rapid biologically-inspired scene classification us-ing features shared with visual attention,”IEEE Transaction on PatternAnalysis and Machine Intelligence, vol. 29, no. 2, pp. 300–312, 2007.
    [133] A. Oliva and M. Greene,“From zero to gist in 200 msec: The time courseof scene recognition,”Scene Understanding Symposium, 2006.
    [134] V. N. Vapnik, Statistical learning theory, Wiley Interscience, New York,1998.
    [135]边肇祺,张学工等,《模式识别》,清华大学出版社,北京, 2000.
    [136] B. Tseng, C.-Y. Lin, M. Naphade, A. Natsev, and J. Smith,“Normalizedclassifier fusion for semantic visual concept detection,”in Proceedings ofthe International Conference on Image Processing, 2003.
    [137] R. Fergus, P. Perona, and A. Zisserman,“A visual category filter for googleimages,”in Proceedings of the 8th European Conference on Computer Vi-sion (ECCV), 2004.
    [138] T. Berg and D. Forsyth,“Animals on the web,”in Proceedings of the IEEEComputer Society Conference on Computer Vision and Pattern Recognition(CVPR), 2006.
    [139] F. Schro?, A. Criminisi, and A. Zisserman,“Harvesting image databasesfrom the web,”in Proceedings of IEEE 11th International Conference onICCV, 2007, pp. 1–8.
    [140] M. Rege, M. Dong, and J. Hua,“Graph theoretical framework for simul-taneously integrating visual and textual features for e?cient web imageclustering,”in Proceeding of the 17th international conference on WorldWide Web, 2008.
    [141] Y. X. Chen, J. Z. Wang, and R. Krovetz,“Clue: cluster-based retrieval ofimages by unsupervised learning,”IEEE Transactions on Image Process-ing, vol. 14, no. 8, pp. 1187–1201, 2005.
    [142] T. Deselaers, D. Keysers, and H. Ney,“Clustering visually similar imagesto improve image search engines,”In Informatiktage 2003 der Gesellschaftfu¨r Informatik, 2003.
    [143] L. Wang, L.J. Yang, and X. M. Tian,“Query aware visual similarity prop-agation for image search reranking,”in Proceedings of ACM Multimedia,2009, pp. 725–728.
    [144] R. van Zwol,“Multimedia strategies for b3-sdr, based on principle compo-nent analysis,”Advances in XML Information Retrieval, Lecture Notes inComputer Science, 2006.
    [145] P. Wilkins, P. Ferguson, and A. F. Smeaton,“Using score distributionsfor query-time fusion in multimedia retrieval,”in Proceedings of the 8thACM international workshop on Multimedia information retrieval, 2006,pp. 51–60.
    [146] R. H. van Leuken, L. G. Pueyo, X. Olivares, and R. van Zwol,“Visualdiversification of image search results,”in Proceedings of WWW, April2009.
    [147] P. Belhumeur, J. Hespanha, and D. Kriegman,“Eigenfaces vs. fisherfaces:Recognition using class specific linear projection,”IEEE Trans. on PatternAnalysis and Machine Intelligence, Special issue on face recognition, vol.19, pp. 711–720, 1997.
    [148] D. Zhang and G. Lu,“Review of shape representation and descriptiontechniques,”Pattern Recognition, vol. 37, no. 1, pp. 1–19, 2004.
    [149]丁险峰等,“形状匹配综述,”自动化学报, vol. 27, no. 5, pp. 678–694, 2001.
    [150] H. Chui and A.Rangarajan,“A new point matching algorithm for non-rigidregistration,”Computer Vision And Image Understanding, vol. 89, no. 2,pp. 114–141, 2003.
    [151] H. Chui and A.Rangarajan,“A new point matching algorithm for non-rigid point matching,”in Proceedings of IEEE International Conferenceon Computer Vision and Pattern Recognition, 2000, pp. 44–51.
    [152] X. Yang, X. Bai, and L. J. Latecki et al,“Improving shape retrieval bylearning graph transduction,”in Proceedings of IEEE ECCV, 2008.
    [153] F. Mahmoudi et al,“Image retrieval based on shape similarity by edgorientation autocorrelogram,”Pattern Recognition, vol. 36, pp. 1725–1736,2003.
    [154] E. L. Kaplan and P. Meier,“Nonparametric estimation from incompleteobservations,”Journal of the American Statistical Association, vol. 53, no.282, pp. 457–481, 1958.
    [155] L. J. Buturovic,“Improving k-nearest neighbor density and error esti-mates,”Pattern Recognition, vol. 26, no. 4, pp. 611–616, 1993.
    [156] B. W. Silverman, Density Estimation for Statistics and Data Analysis,Chapman and Hall, 1986.
    [157] M. Rosenblatt,“Remarks on some nonparametric estimates of a densityfunction,”Annals of Mathematical statistics, vol. 27, no. 6, pp. 832–837,1956.
    [158] E. Parzen,“On estimation of a probability density function and mode,”Annals of Mathematical statistics, vol. 33, no. 3, pp. 1065–1076, 1962.
    [159] N. Dalal and B. Triggs,“Histograms of oriented gradients for human de-tection,”in Proceedings of the IEEE Computer Society Conference onComputer Vision and Pattern Recognition (CVPR), 2005.
    [160] R. Zabih and J. Woodfill,“Non-parametric local transforms for computingvisual correspondence,”in Proceedings of ECCV, 1994, pp. 151–158.

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700