- -

HumanTop: a multi-object tracking tabletop

RiuNet: Repositorio Institucional de la Universidad Politécnica de Valencia

Compartir/Enviar a

Citas

Estadísticas

  • Estadisticas de Uso

HumanTop: a multi-object tracking tabletop

Mostrar el registro sencillo del ítem

Ficheros en el ítem

dc.contributor.author Soto Candela, Emilio es_ES
dc.contributor.author Ortega Pérez, Mario es_ES
dc.contributor.author Marín Romero, Clemente es_ES
dc.contributor.author Pérez López, David Clemente es_ES
dc.contributor.author Salvador Herranz, Gustavo Manuel es_ES
dc.contributor.author Contero, Manuel es_ES
dc.contributor.author Alcañiz Raya, Mariano Luis es_ES
dc.date.accessioned 2017-06-09T10:13:41Z
dc.date.available 2017-06-09T10:13:41Z
dc.date.issued 2014-06
dc.identifier.issn 1380-7501
dc.identifier.uri http://hdl.handle.net/10251/82640
dc.description.abstract In this paper, a computer vision based interactive multi-touch tabletop system called HumanTop is introduced. HumanTop implements a stereo camera vision subsystem which allows not only an accurate fingertip tracking algorithm but also a precise touch-over-the-working surface detection method. Based on a pair of visible spectra cameras, a novel synchronization circuit makes the camera caption and the image projection independent from each other, providing the minimum basis for the development of computer vision analysis based on visible spectrum cameras without any interference coming from the projector. The assembly of both cameras and the synchronization circuit is not only capable of performing an ad-hoc version of a depth camera, but it also introduces the recognition and tracking of textured planar objects, even when contents are projected over them. On the other hand HumanTop supports the tracking of sheets of paper and ID-code markers. This set of features makes the HumanTop a comprehensive, intuitive and versatile augmented tabletop that provides multitouch interaction with projective augmented reality on any flat surface. As an example to exploit all the capabilities of HumanTop, an educational application has been developed using an augmented book as a launcher to different didactic contents. A pilot study in which 28 fifth graders participated is presented. Results about efficiency, usability/satisfaction and motivation are provided. These results suggest that HumanTop is an interesting platform for the development of educational contents. © 2012 Springer Science+Business Media, LLC. es_ES
dc.description.sponsorship This study was funded by Ministerio de Educacion y Ciencia Spain, Project SALTET (TIN2010-21296-C02-01), Project Game Teen (TIN2010-20187) projects Consolider-C (SEJ2006-14301/PSIC), "CIBER of Physiopathology of Obesity and Nutrition, an initiative of ISCIII" and Excellence Research Program PROMETEO (Generalitat Valenciana. Conselleria de Educacio, 2008-157). en_EN
dc.language Inglés es_ES
dc.publisher Springer Verlag (Germany) es_ES
dc.relation.ispartof Multimedia Tools and Applications es_ES
dc.rights Reserva de todos los derechos es_ES
dc.subject Camera-projector system es_ES
dc.subject Finger detection es_ES
dc.subject Markerless tracking es_ES
dc.subject Multitouch es_ES
dc.subject Tabletop es_ES
dc.subject Technology enhanced learning es_ES
dc.subject.classification EXPRESION GRAFICA EN LA INGENIERIA es_ES
dc.title HumanTop: a multi-object tracking tabletop es_ES
dc.type Artículo es_ES
dc.identifier.doi 10.1007/s11042-012-1193-y
dc.relation.projectID info:eu-repo/grantAgreement/MICINN//TIN2010-21296-C02-01/ES/HERRAMIENTAS TECNOLOGICAS AVANZADAS PARA APOYO AL APRENDIZAJE ACTIVO Y ENTRENAMIENTO TECNICO/ es_ES
dc.relation.projectID info:eu-repo/grantAgreement/MICINN//TIN2010-20187/ES/ENTORNOS INMERSIVOS Y PERSUASIVOS PARA LA EVALUACION Y ENTRENAMIENTO DE ESTRATEGIAS DE REGULACION EMOCIONAL. APLICACION A LA EDUCACION PSICOSOCIAL EN ADOLESCENTES/ es_ES
dc.relation.projectID info:eu-repo/grantAgreement/MEC//SEJ2006-14301/ES/NUEVAS TECNOLOGIAS DE LA INFORMACION Y LA COMUNICACION: INTEGRACION Y CONSOLIDACION DE SU USO EN CIENCIAS SOCIALES PARA MEJORAR LA SALUD, LA CALIDAD DE VIDA Y EL BIENESTAR./ es_ES
dc.relation.projectID info:eu-repo/grantAgreement/GVA//PROMETEO08%2F2008%2F157/ES/Promoción del bienestar a través de las tecnologías de la información y comunicación (probientic)/ es_ES
dc.rights.accessRights Cerrado es_ES
dc.contributor.affiliation Universitat Politècnica de València. Escuela Técnica Superior de Ingenieros Industriales - Escola Tècnica Superior d'Enginyers Industrials es_ES
dc.contributor.affiliation Universitat Politècnica de València. Instituto Interuniversitario de Investigación en Bioingeniería y Tecnología Orientada al Ser Humano - Institut Interuniversitari d'Investigació en Bioenginyeria i Tecnologia Orientada a l'Ésser Humà es_ES
dc.contributor.affiliation Universitat Politècnica de València. Escuela Técnica Superior de Ingeniería Agronómica y del Medio Natural - Escola Tècnica Superior d'Enginyeria Agronòmica i del Medi Natural es_ES
dc.contributor.affiliation Universitat Politècnica de València. Escuela Politécnica Superior de Alcoy - Escola Politècnica Superior d'Alcoi es_ES
dc.description.bibliographicCitation Soto Candela, E.; Ortega Pérez, M.; Marín Romero, C.; Pérez López, DC.; Salvador Herranz, GM.; Contero, M.; Alcañiz Raya, ML. (2014). HumanTop: a multi-object tracking tabletop. Multimedia Tools and Applications. 70(3):1837-1868. https://doi.org/10.1007/s11042-012-1193-y es_ES
dc.description.accrualMethod S es_ES
dc.relation.publisherversion http://dx.doi.org/10.1007/s11042-012-1193-y es_ES
dc.description.upvformatpinicio 1837 es_ES
dc.description.upvformatpfin 1868 es_ES
dc.type.version info:eu-repo/semantics/publishedVersion es_ES
dc.description.volume 70 es_ES
dc.description.issue 3 es_ES
dc.relation.senia 230712 es_ES
dc.identifier.eissn 1573-7721
dc.contributor.funder Ministerio de Educación y Ciencia es_ES
dc.contributor.funder Generalitat Valenciana es_ES
dc.description.references Agarwal A, Izadi S, Chandraker M, Blake A (2007) High precision multi-touch sensing on surfaces using overhead cameras. In: IEEE int. workshop horiz. interact. hum.-comput. interact., TABLETOP’07. IEEE, pp 197–200 es_ES
dc.description.references Alexa M, Bollensdorff B, Bressler I, Elstner S, Hahne U, Kettlitz N, Lindow N, Lubkoll R, Richter R, Stripf C et al (2008) Continuous reference images for ftir touch sensing. In: ACM SIGGRAPH poster. ACM, p 49 es_ES
dc.description.references Argyros A, Lourakis M (2006) Vision-based interpretation of hand gestures for remote control of a computer mouse. In: Comput. vis. hum.-comput. interact., pp 40–51 es_ES
dc.description.references Barnes C, Jacobs D, Sanders J, Goldman D, Rusinkiewicz S, Finkelstein A, Agrawala M (2008) Video puppetry: a performative interface for cutout animation. ACM Trans Graph (TOG) 27:124 es_ES
dc.description.references Bradski G, Kaehler A (2008) Learning OpenCV: computer vision with the OpenCV library. O’Reilly Media es_ES
dc.description.references Campbell D, Stanley J, Gage N (1963) Experimental and quasi-experimental designs for research. Houghton Mifflin, Boston es_ES
dc.description.references Chen D, Zhang G (2005) A new sub-pixel detector for x-corners in camera calibration targets. In: 13th int. conf. cent. Eur. comput. graph., vis. comput. vis. es_ES
dc.description.references Dietz P, Leigh D (2001) Diamondtouch: a multi-user touch technology. In: Proc. 14th ACM symp. user interface softw. technol. ACM, pp 219–226 es_ES
dc.description.references Do-Lenh S, Kaplan F, Sharma A, Dillenbourg P (2009) Multi-finger interactions with papers on augmented tabletops. In: Proc. 3rd int. conf. tangible embed. int. ACM, pp 267–274 es_ES
dc.description.references Dung L, Mizukawa M (2009) Fast hand feature extraction based on connected component labeling, distance transform and hough transform. J. Robot. Mechatronics 21(6):726–738 es_ES
dc.description.references Echtler F, Sielhorst T, Huber M, Klinker G (2009) A short guide to modulated light. In: Proc. 3rd int. conf. tang. embed. interact. ACM, pp 393–396 es_ES
dc.description.references Echtler F, Pototschnig T, Klinker G (2010) An led-based multitouch sensor for lcd screens. In: Proc. 4th int. conf. tang. embed. interact.. ACM, pp 227–230 es_ES
dc.description.references Han J (2005) Low-cost multi-touch sensing through frustrated total internal reflection. In: Proc. 18th ACM symp. user interface softw. technol. ACM, pp 115–118 es_ES
dc.description.references Holman D, Vertegaal R, Altosaar M, Troje N, Johns D (2005) Paper windows: interaction techniques for digital paper. In: Proc. SIGCHI conf. hum. factor comput. syst. ACM, pp 591–599 es_ES
dc.description.references Izadi S, Agarwal A, Criminisi A, Winn J, Blake A, Fitzgibbon A (2007) C-slate: a multi-touch and object recognition system for remote collaboration using horizontal surfaces. In: IEEE int. workshop horiz. interact. hum.-comput. interact., TABLETOP’07. IEEE, pp 3–10 es_ES
dc.description.references Jordà S, Geiger G, Alonso M, Kaltenbrunner M (2007) The reactable: exploring the synergy between live music performance and tabletop tangible interfaces. In: Proc. 1st int. conf. tangible embed. interact. ACM, pp 139–146 es_ES
dc.description.references Kaltenbrunner M (2009) Reactivision and tuio: a tangible tabletop toolkit. In: Proc. ACM int. conf. interact. tabletop. surf. ACM, pp 9–16 es_ES
dc.description.references Katz I, Gabayan K, Aghajan H (2007) A multi-touch surface using multiple cameras. In: Proc. 9th int. conf. adv. concept. intell. vis. syst.. Springer, pp 97–108 es_ES
dc.description.references Kim K, Lepetit V, Woo W (2010) Scalable real-time planar targets tracking for digilog books. Vis Comput 26(6):1145–1154 es_ES
dc.description.references Lee T, Hollerer T (2007) Handy ar: markerless inspection of augmented reality objects using fingertip tracking. In: 11th IEEE int. symp. wearable comput. IEEE, pp 83–90 es_ES
dc.description.references Letessier J, Bérard F (2004) Visual tracking of bare fingers for interactive surfaces. In: Proc. 17th ACM symp. user interface softw. technol. ACM, pp 119–122 es_ES
dc.description.references Likert R (1932) A technique for the measurement of attitudes. Arch Psychol 140:1–55 es_ES
dc.description.references Lucchese L, Mitra S (2002) Using saddle points for subpixel feature detection in camera calibration targets. In: Asian-Pac. conf. circuit. syst., vol 2. IEEE, pp 191–195 es_ES
dc.description.references Malik S, Laszlo J (2004) Visual touchpad: a two-handed gestural input device. In: Proc. 6th int. conf. multimodal interface. ACM, pp 289–296 es_ES
dc.description.references Manresa C, Varona J, Mas R, Perales F (2000) Real–time hand tracking and gesture recognition for human-computer interaction. In: Comput. vis. cent., pp 1–7 es_ES
dc.description.references Martín-Gutiérrez J, Luís Saorín J, Contero M, Alcañiz M, Pérez-López D, Ortega M (2010) Design and validation of an augmented book for spatial abilities development in engineering students. Comput Graph 34(1):77–91 es_ES
dc.description.references McNaughton J (2010) Utilising emerging multi-touch table designs. Durham University es_ES
dc.description.references Microsoft (2011) Microsoft surface. URL http://www.microsoft.com/surface/ es_ES
dc.description.references Muja M, Lowe D (2009) Fast approximate nearest neighbors with automatic algorithm configuration. In: Int. conf. comput. vis. theory appl. VISSAPP, pp 331–340 es_ES
dc.description.references Nister D, Stewenius H (2006) Scalable recognition with a vocabulary tree. In: IEEE Comput. Soc. conf. comput. vis. pattern recognit., vol 2. IEEE, pp 2161–2168 es_ES
dc.description.references Oka K, Sato Y, Koike H (2002) Real-time fingertip tracking and gesture recognition. IEEE Comput Graph 22(6):64–71 es_ES
dc.description.references OpenSource (2011) Fast sift image features library. URL http://libsift.sourceforge.net/ es_ES
dc.description.references Peer P, Kovac J, Solina F (2003) Human skin color clustering for face detection, vol 2. IEEE es_ES
dc.description.references Pilet J, Saito H (2010) Virtually augmenting hundreds of real pictures: an approach based on learning, retrieval, and tracking. In: IEEE virtual real. conf. (VR). IEEE, pp 71–78 es_ES
dc.description.references Rekimoto J (2002) Smartskin: an infrastructure for freehand manipulation on interactive surfaces. In: Proc. SIGCHI conf. hum. factor. comput. syst.. ACM, pp 113–120 es_ES
dc.description.references Shi J, Tomasi C (1994) Good features to track. In: IEEE comput. soc. conf. proc. comput. vis. pattern recognit. IEEE, pp 593–600 es_ES
dc.description.references Tomasi C, Kanade T (1991) Detection and tracking of point features. School of Computer Science, Carnegie Mellon University es_ES
dc.description.references Verdié Y (2008) Evolution of hand tracking algorithms to mirrortrack. Tech. Rep. Vis. Interface Syst. Lab. es_ES
dc.description.references Vos N, van der Meijden H, Denessen E (2011) Effects of constructing versus playing an educational game on student motivation and deep learning strategy use. Comput Educ 56(1):127–137 es_ES
dc.description.references Wagner D, Reitmayr G, Mulloni A, Drummond T, Schmalstieg D (2010) Real-time detection and tracking for augmented reality on mobile phones. IEEE Trans Vis Comput Graph 16(3):355–368 es_ES
dc.description.references Welch G, Bishop G (1995) An introduction to the Kalman filter. University of North Carolina at Chapel Hill, Citeseer es_ES
dc.description.references Wilson A (2004) Touchlight: an imaging touch screen and display for gesture-based interaction. In: Proc. 6th int. conf. multimodal interface. ACM, pp 69–76 es_ES
dc.description.references Wilson A (2005) Playanywhere: a compact interactive tabletop projection-vision system. In: Proc. 18th ACM symp user interface softw. technol. ACM, pp 83–92 es_ES
dc.description.references Wilson A (2010) Using a depth camera as a touch sensor. In: ACM int. conf. interact. tabletop. surf. ACM, pp 69–72 es_ES
dc.description.references Zerofrog (2011) Libsiftfast. URL http://sourceforge.net/projects/libsift es_ES
dc.description.references Zhang Z (2000) A flexible new technique for camera calibration. IEEE Trans Pattern Anal Mach Intell 22(11):1330–1334 es_ES
dc.description.references Zhang Z, Wu Y, Shan Y, Shafer S (2001) Visual panel: virtual mouse, keyboard and 3d controller with an ordinary piece of paper. In: Proc. workshop percept. user interface. ACM, pp 1–8 es_ES


Este ítem aparece en la(s) siguiente(s) colección(ones)

Mostrar el registro sencillo del ítem