







Vol.4 , No. 1, Publication Date: Jun. 9, 2017, Page: 1-15
[1] | Alessandra Lumini, Department of Computer Science and Engineering, University of Bologna, Cesena, Italy. |
This paper presents a novel mobile application for the implementation on Android mobile devices of a photo treasure hunt. GeoPhotoHunt (GPH) provides services for creating/playing a photographic treasure hunt as a guide for city visiting or for recreational purposes. The application uses computer vision algorithms executed onboard to quantify similarity between the picture captured by the phone camera and the correct image, without requiring a remote server and an internet connection: this is a main requirement for tourists who often pay a lot of money for international data roaming. Both creation and playing phases are performed on the mobile device in order to facilitate the use on site. The image recognition module on the mobile device manages the feature extraction and the calculation of similarity to the stored objects. In this paper we report a rather broad analysis of existing works in the domains of location-based pervasive games applications and image recognition using limited resources. Moreover, in order to select the best image similarity method, we perform a wide comparison of descriptors used to represent images and measures used to evaluate their similarity. Our experiments carried out on 7 image datasets prove that the system is efficient and robust in comparing images with different characteristics.
Keywords
Photographic Treasure Hunt, Instance Recognition, Computer Vision, Image Content Analysis, Mobile CBIR
Reference
[01] | S. Benford, C. Magerkurth, and P. Ljungstrand, “Bridging the physical and digital in pervasive gaming,” Communications of the ACM, vol. 48, no. 3. p. 54, 2005. |
[02] | D. Nicklas, C. Pfisterer, and B. Mitschang, “Towards location-based games,” in Proceedings of the International Conference on Applications and Development of Computer Games in the 21st Century: ADCOG, 2001, vol. 21, pp. 61–67. |
[03] | R. M. Webb, “Recreational geocaching: the Southeast Queensland experience,” 2001. |
[04] | G. Amato, F. Falchi, and F. Rabitti, “Landmark recognition in VISITO Tuscany,” in Communications in Computer and Information Science, 2012, vol. 247 CCIS, pp. 1–13. |
[05] | C. Casanova, A. Franco, A. Lumini, and D. Maio, “SmartVisionApp: A framework for computer vision applications on mobile devices,” Expert Syst. Appl., vol. 40, no. 15, pp. 5884–5894, Nov. 2013. |
[06] | G. Vavoula, M. Sharples, P. Rudman, J. Meek, and P. Lonsdale, “Myartspace: Design and evaluation of support for learning with multimedia phones between classrooms and museums,” Comput. Educ., vol. 53, no. 2, pp. 286–299, 2009. |
[07] | A. Giemza, N. Malzahn, and H. U. Hoppe, “Mobilogue: Creating and Conducting Mobile Learning Scenarios in Informal Settings,” in The 21st International Conference on Computers in Education (ICCE 2013), 2013. |
[08] | L. Botturi, A. Inversini, and A. Di Maria, “The City Treasure: Mobile Games for Learning Cultural Heritage,” in Museums and the Web 2010: Proceedings, 2009. |
[09] | D. Kohen-Vacs, M. Ronen, and S. Cohen, “Mobile Treasure Hunt Games for Outdoor Learning,” Bull. IEEE Tech. Comm. Learn. Technol., vol. 14, no. 4, pp. 24–26, 2012. |
[10] | D. Grüntjens, S. Groß, D. Arndt, and S. Müller, “Fast authoring for mobile gamebased city tours,” in Procedia Computer Science, 2013, vol. 25, pp. 41–51. |
[11] | T. J. Chin, Y. You, C. Coutrix, J. H. Lim, J. P. Chevallet, and L. Nigay, “Mobile phone-based mixed reality: The Snap2Play game,” Vis. Comput., vol. 25, no. 1, pp. 25–37, 2009. |
[12] | P. Bahl and V. N. Padmanabhan, “RADAR: an in-building RF-based user location and tracking system,” Proc. IEEE INFOCOM 2000. Conf. Comput. Commun. Ninet. Annu. Jt. Conf. IEEE Comput. Commun. Soc. (Cat. No.00CH37064), vol. 2, 2000. |
[13] | P. Föckler, T. Zeidler, B. Brombach, E. Bruns, and O. Bimber, “PhoneGuide: museum guidance supported by on-device object recognition on mobile phones,” Proc. 4th Int. Conf. Mob. ubiquitous Multimed., vol. 6, pp. 3–10, 2005. |
[14] | B. Ruf, E. Kokiopoulou, and M. Detyniecki, “Mobile museum guide based on fast SIFT recognition,” in Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2010, vol. 5811 LNCS, pp. 170–183. |
[15] | A. Mody, M. Akram, K. Rony, S. A. Muhammad, and R. Kamoua, “Enhancing user experience at museums using smart phones with RFID,” in 2009 IEEE Long Island Systems, Applications and Technology Conference, LISAT 2009, 2009. |
[16] | V. Tyagi, A. S. Pandya, A. Agarwal, and B. Alhalabi, “Validation of object recognition framework on Android mobile platform,” in Proceedings of IEEE International Symposium on High Assurance Systems Engineering, 2011, pp. 313–316. |
[17] | T. Yeh, K. Tollmar, and T. Darrell, “Searching the Web with mobile images for location recognition,” Proc. 2004 IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognition, 2004. CVPR 2004., vol. 2, 2004. |
[18] | G. Fritz, C. Seifert, M. Kumar, and L. Paletta, “Building Detection from Mobile Imagery Using Informative SIFT Descriptors,” in Scandinavian Conference on Image Analysis, 2005, pp. 629–638. |
[19] | M. F. Arriaga-Gomez, I. de Mendizabal-Vazquez, R. Ros-Gomez, and C. Sanchez-Avila, “A comparative survey on supervised classifiers for face recognition,” in Security Technology (ICCST), 2014 International Carnahan Conference on, 2014, pp. 1–6. |
[20] | K. H. Yap, T. Chen, Z. Li, and K. Wu, “A comparative study of mobile-based landmark recognition techniques,” in IEEE Intelligent Systems, 2010, vol. 25, no. 1, pp. 48–57. |
[21] | P. Bhattacharya and M. Gavrilova, “A survey of landmark recognition using the bag-of-words framework,” in Intelligent computer graphics 2012, Springer, 2013, pp. 243–263. |
[22] | A. Redondi, M. Cesana, and M. Tagliasacchi, “Low bitrate coding schemes for local image descriptors.,” in MMSP, 2012, pp. 124–129. |
[23] | V. Hedau, S. N. Sinha, C. L. Zitnick, and R. Szeliski, “A Memory Efficient Discriminative Approach for Location Aided Recognition.,” in ECCV Workshops (1), 2012, vol. 7583, pp. 187–197. |
[24] | G. Takacs, V. Chandrasekhar, N. Gelfand, Y. Xiong, W.-C. Chen, T. Bismpigiannis, R. Grzeszczuk, K. Pulli, and B. Girod, “Outdoors augmented reality on mobile phone using loxel-based visual feature organization,” in MIR ’08: Proceeding of the 1st ACM international conference on Multimedia information retrieval, 2008, pp. 427–434. |
[25] | H. Bay, A. Ess, T. Tuytelaars, and L. Van Gool, “Speeded-Up Robust Features (SURF),” Comput. Vis. Image Underst., vol. 110, no. 3, pp. 346–359, 2008. |
[26] | K. Lin, J. Li, W. Huang, and C. Fang, “Smartphone landmark image retrieval based on Lucene and GPS,” in Computer Science Education (ICCSE), 2014 9th International Conference on, 2014, pp. 368–371. |
[27] | X. Yang and K. T. Cheng, “Learning optimized local difference binaries for scalable augmented reality on mobile devices,” IEEE Trans. Vis. Comput. Graph., vol. 20, no. 6, pp. 852–865, 2014. |
[28] | C. W. Niblack, “QBIC project: querying images by content, using color, texture, and shape,” Proc. SPIE, vol. 1908, no. 1, pp. 173–187, 1993. |
[29] | W. Y. Ma, Y. Deng, and B. S. Manjunath, “Tools for texture/color based search of images,” Hum. Vis. Electron. Imaging II, vol. 3016, pp. 496–507, 1997. |
[30] | O. A. B. Penatti, E. Valle, and R. da S. Torres, “Comparative study of global color and texture descriptors for web image retrieval,” J. Vis. Commun. Image Represent., vol. 23, no. 2, pp. 359–380, 2012. |
[31] | M. J. Swain and D. H. Ballard, “Color indexing,” Int. J. Comput. Vis., vol. 7, no. 1, pp. 11–32, 1991. |
[32] | M. Stricker and M. Orengo, “Similarity of color images,” in Proc. SPIE Storage and Retrieval for Image and Video Databases, 1995, vol. 2420, pp. 381–392. |
[33] | R. O. Stehling, M. A. Nascimento, and A. X. Falcao, “An adaptive and efficient clustering-based approach forcontent-based image retrieval in image databases,” Proc. 2001 Int. Database Eng. Appl. Symp., 2001. |
[34] | R. M. Haralick, “STATISTICAL AND STRUCTURAL APPROACHES TO TEXTURE.,” Proc IEEE, vol. 67, no. 5. pp. 786–804, 1979. |
[35] | R. Nevatia, Machine Perception. Prentice-Hall, 1982. |
[36] | T. Ojala, M. Pietikainen, and T. Maenpaa, “Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 24, no. 7, pp. 971–987, 2002. |
[37] | P. Wu, B. S. Manjunath, S. Newsam, and H. D. Shin, “Texture descriptor for browsing and similarity retrieval,” Signal Process. Image Commun., vol. 16, no. 1, pp. 33–43, 2000. |
[38] | N. Dalal and W. Triggs, “Histograms of Oriented Gradients for Human Detection,” 2005 IEEE Comput. Soc. Conf. Comput. Vis. Pattern Recognit. CVPR05, vol. 1, no. 3, pp. 886–893, 2004. |
[39] | a Khotanzad and Y. H. Hong, “Invariant image recognition by Zernike moments,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 12, no. 5, pp. 489–497, 1990. |
[40] | D. G. Lowe, “Distinctive image features from scale-invariant keypoints,” Int. J. Comput. Vis., vol. 60, pp. 91–110, 2004. |
[41] | E. Rublee, V. Rabaud, K. Konolige, and G. Bradski, “ORB: An efficient alternative to SIFT or SURF,” in Proceedings of the IEEE International Conference on Computer Vision, 2011, pp. 2564–2571. |
[42] | G. Csurka, C. R. Dance, L. Fan, J. Willamowski, and C. Bray, “Visual categorization with bags of keypoints,” in Proceedings of the ECCV International Workshop on Statistical Learning in Computer Vision, 2004, pp. 59–74. |
[43] | G. Pass, R. Zabih, and J. Miller, “Comparing images using color coherence vectors,” in Proceedings of the fourth ACM international conference on Multimedia (MULTIMEDIA ’96), 1996, pp. 65–73. |
[44] | X. Niu and Y. Jiao, “An overview of perceptual hashing,” Acta Electron. Sin., vol. 36, no. 7, pp. 1405–1411, 2008. |
[45] | T. Ojala, M. Pietikäinen, and T. Mäenpää, “Multiresolution gray-scale and rotation invariant texture classification with local binary patterns,” \TPAMI, vol. 24, no. 7, pp. 971–987, 2002. |
[46] | C. Beecks, M. Seran Uysal, and T. Seidl, “A comparative study of similarity measures for content-based multimedia retrieval,” in 2010 IEEE International Conference on Multimedia and Expo, ICME 2010, 2010, pp. 1552–1557. |
[47] | M. Aly, P. Welinder, M. Munich, and P. Perona, “Towards automated large scale discovery of image families,” in 2009 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2009, 2009, pp. 9–16. |
[48] | H. S. H. Shao, T. Svoboda, V. Ferrari, T. Tuytelaars, and L. Van Gool, “Fast indexing for image retrieval based on local appearance with re-ranking,” Proc. 2003 Int. Conf. Image Process. (Cat. No.03CH37429), vol. 3, 2003. |
[49] | J. Li and N. M. Allinson, “Subspace learning-based dimensionality reduction in building recognition,” Neurocomputing, vol. 73, no. 1–3, pp. 324–330, Dec. 2009. |
[50] | C. X. Ling, J. Huang, and H. Zhang, “AUC: A better measure than accuracy in comparing learning algorithms,” Adv. Artif. Intell., pp. 329–341, 2003. |