Keynote slides
Transcription
Keynote slides
ECCV WORKSHOP VISART ZURICH SEPTEMBER 2014 Computer Vision for interactive experiences with art and artistic documents Prof. Rita Cucchiara DIPARTIMENTO DI INGEGNERIA Enzo Ferrari Università di Modena e Reggio Emilia, Italia http://www.Imagelab.ing.unimore.it ABOUT OUR WORK IMAGELAB imagelab.ing@unimore.it SOFTECH-ICT research center in ICT for enterprise Dipartimento di Ingegneria Enzo Ferrari Research in computer vision , pattern recognition Multimedia and machine learning - Vision for Video Surveillance (since 1999) - Vision for Medical imaging - Vision for Industry - Vision for cultural experiences Rita Cucchiara ECCV W VISART 2014, Italy Università di Modena e Reggio Emilia, ITALY Open post-doc positions!! Please contact me CULTURAL HERITAGE UNESCO 2003 cultural heritage definitions: • tangible heritage : artifacts such as objects, paintings, buildings, structures, landscapes, cities. (cultural heritage, natural heritage) • intangible heritage : practices, representations, expressions, memories, knowledge that are naturally transmitted in oral or written forms. Modena, Italy The Dolomiti, Italy The Mediterrean diet, Italy • digital heritage is made up of computer-based materials of enduring value that should be kept for future generations: all multimedia data texts, images, audio, graphics, software and web pages, native digital or after digitalization. Rita Cucchiara ECCV W VISART 2014, Italy COMPUTER VISION FOR INTERACTIVITY Interacting with the Great Beauty: 1)Computer vision for human-centered (digital) activities in experience with art and cultural heritage ( monitoring, reconstruction AR, retrieval, learning and understanding,) 2)Computer vision for human augmentation to improve the experience with art and cultural heritage ( visual augmentation, natural HCI, enjoying) Rita Cucchiara ECCV W VISART 2014, Italy COMPUTER VISION FOR INTERACTIVITY Computer vision for interactivity with art is not (only) interactive art. but they started together… Myron Krueger’s Videoplace, (1975 for “artificial reality”, SIGGRAPH 1985)…. Before mouse! Rita Cucchiara ECCV W VISART 2014, Italy CV FOR INTERACTING WITH CH Envi-Vision: Vision by environment, fixed or moving cameras Ego-vision: Vision by mobile, wearable egocentric, cameras 1)for seeing what your eyes don’t reach 2)for seeing what your eyes cannot see 3)for telling what your eyes are seeing 4) for seeing with more eyes Rita Cucchiara, Alberto Del Bimbo, "Visions for Augmented Cultural Heritage Experience," IEEE Multimedia, vol. 21, no. 1, pp. 74-82, Jan.-Mar., 2014. Rita Cucchiara ECCV W VISART 2014, Italy 1.FOR SEEING WHAT YOUR EYES DON’T REACH • Webcams • Surveillance cameras mounted in museums and CH locations for seeing interactions Italy: Project Cluster SC&C MNEMOSYNE [Baldanov et al MCH Workshop 2012] • Drones • Images for 3D reconstruction [Pollefeys, Van Gool ACM Surv 2003] The Great Buddha project 02-07 [Hiazaki ..Ikeuki JCV 2007] Rita Cucchiara ECCV W VISART 2014, Italy 2. FOR SEEING WHAT YOUR EYES CANNOT SEE • Thermal cameras for monitoring • Floor cameras for interactions…* FLORIMAGE project, LECCE Museum 2015 • Stereo and range scanners for 3D reconstruction • Deep image processing as David does… *M. Lombardi, A. Pieracci, P. Santinelli, R. Vezzani, R. Cucchiara,"Human Behavior Understanding with Wide Area Sensing Floors" in HBU2013, vol. 8212, LNCS 8212, pp. 112-123, 2013 ITALIAN Project PON DICET, Lecce 2013-2016 Rita Cucchiara ECCV W VISART 2014, Italy 3. FOR TELLING WHAT YOUR EYES ARE SEEING Typical interaction from mobile vision (Google Goggles) Augmented reality and vision [Caarls et al JIVP 2009] • Gaming Experiments for 3D retrieval by mobile (Enzo Ferrari Museum 2013) • Vision and augmented reality in cultural and natural sites (MARMOTA FBK: FP7 VENTURI ) • Document and painting recognition for retrieval • As James and many of us Rita Cucchiara ECCV W VISART 2014, Italy DOCUMENT RECOGNITION FOR RETRIEVAL Computer vision CV HCI - Image Retrieval [Zhang PAMI2012] - Image segmentation and multi-digitalization - Multi-digitalization of minate code manuscripts - - The DE RERUM NOVARUM Project + Multi-digitalization of Digitalized Encyclopedia - The Treccani-DICET project** +D. Borghesani, C. Grana, R. Cucchiara, "Miniature illustrations retrieval and innovative interaction for digital illuminated manuscripts"in Multimedia Systems, 2013 **D.Coppi, C.Grana, R. Cucchiara Illustration Segmentationin in Digitalized documents using local correlation features Proc. of ICRDL 2014, MTAP to appear Rita Cucchiara ECCV W VISART 2014, Italy DOCUMENT ANALYSIS FOR INTERACTION Multi-digitalization of artistic books for new forms of digital interactivities - Layout segmentation - Picture segmentation and tagging - Copy detection - Search with relevant feedback web intera ction Manual annotation Document Analysis Digitalization Papery Documents Rita Cucchiara ECCV W VISART 2014, Italy Digital Library Mutimedia interactive Digital Library CV FOR AUTOMATIC PICTURE SELECTION 1400 AC The Holy Bible of Borso D’Este Content-based image retrieval Backgrou nd Text User interface Image Picture Decoratio n Feature annotatio n Thanks to D.BoRghesani,C.Grana 2010-13 Rita Cucchiara ECCV W VISART 2014, Italy FEATURES CORRELATION MATRIX FEATURE POINTS FROM REIMANNIAN TO EUCLIDEAN SPACE AUTOCORRELATION DIRECTIONAL HISTOGRAMS Rita Cucchiara ECCV W VISART 2014, Italy [Borghesani et al ACM J Multimedia Systems 2014] [Borghesani et al MTAP 2012] Segmentation and tagging on digitalized books ACM Multimedia 2010, MTAP2011 Interactive surfing on digitalized books Adding relevance feedback by users Adding positive and negative relevance feedback Improving search by similarity After multi-digitalization a Multitouch Interface 2012 Software written in C++ using Nokia QT4 libraries Supports Windows, Mac and Linux 46’’ LCD panel equipped with a 32 point multitouch Now a multidigitalized product INTERACTION WITH ENCYCLOPEDIA Multi-digitalization of Treccani Encyclopedia Blocks feature extraction Training model SVM classification XY Cut Blocks Autocorrelation Image segmentation and block classification nn blocks extraction Rita Cucchiara ECCV W VISART 2014, Italy Thanks to C.Grana, M.Fornaciari, D.Coppi INTERACTING WITH ENCYCLOPEDIA Layour segmentation Specific tailored for drawings and artistic schemes Tesseract Rita Cucchiara ECCV W VISART 2014, Italy Our Method Courtesy of Treccani Enciclopedia AUTOCORRELATION MATRIX • Block analysis • Visual features extraction • Feature classification Dataset Treccani Gutenberg13 Method Our Tesseract Our Tesseract % TP % FN % FP 99,57 0,43 4,53 52,25 47,71 0,39 99,50 0,50 11,50 83,13 16,87 1,02 Rita Cucchiara ECCV W VISART 2014, Italy After multi-digitalization: Web retrieval Searching and interacting with Treccani images through the web GOLD descriptors ( Gaussian of Local Descriptors) (best accuracy at IMAGECLEF2013) C.Grana, G. Serra, M. Manfredi, R. Cucchiara,"Beyond Bag of Words for Concept Detection and Search of Cultural Heritage Archives« in SISAP 2013, vol. 8199, LNCS 8199, Spain, pp. 233-244, Oct. 2-4, 2013 TUTORIAL AT ICPR2014 Rita Cucchiara ECCV W VISART 2014, Italy Thanks to C.Grana, G.Serra, M. Manfredi 4. FOR SEEING WITH MORE EYES Physical, augmented experience with tangible CH The explosion of wearable cameras and ego-centric vision: - wearable museum or city guides - self gesture analysis for interactions - real-time recognition of CH targets Using Vision for detecting/tracking/recognizing targets and observers’ interaction Rita Cucchiara ECCV W VISART 2014, Italy EGOCENTRIC VISION Egocentric vision ( “EgoVision”) models and techniques for understanding what a person sees, from the first person’s point of view and centered on the human perceptual needs. Often called first-person vision, to recall the needs of using wearable cameras (e.g. on glasses mounted on the head) for acquiring and processing the same visual stimuli that human acquire and process. a broader meaning ….. to understand what a person sees or want to see and to exploit similar learning, perception and reasoning paradigms of humans.. Rita Cucchiara ECCV W VISART 2014, Italy CV CHALLENGES FOR CULTURAL EXPERIENCE IN EGO-VISION Cultural experience in ego-vision Life Logging Organizing memory and data Off-line Big data Deep learning Storage and transmission issues Rita Cucchiara ECCV W VISART 2014, Italy Human Augmentation Understanding world by vision… On-line Noisy, unconstrained data Fast learning with few examples Processing issues A CULTURAL EXPERIENCE WITH EGOVISION & Computer vision Rita Cucchiara ECCV W VISART 2014, Italy Thanks to Giuseppe Serra, Stefano Aletto, Lorenzo Baldini Francesco Paci, Luca Benini @ETHZ CHALLENGES IN EGO-VISION Egovision for visual augmentation … * Similar to video-surveillance and robot vision • • • • fast, real-time ( please, limit the data searchspace ) reactive pro-active similar scenes ( typically people, social life, children..) many similar methods ( detection, action recognition , tracking) but • Unconstrained • Large different motion factors • Frequent Changes of field of view • Very Long videos * R.Cucchiara Egocentric vision tarcking and evaluating human signs ICVSS Catania 2014 Rita Cucchiara ECCV W VISART 2014, Italy CV CHALLENGES IN EGOVISION(1/3) 1. Hardware • Design new hardware • Exploit real-time capabilities for egovision 2. Recognizing FoA /PoI • Estimating FoA[Li ICCV2013], [OgakiCVPRW2012], [Jianfeng CVPR2014] • Eye-tracking; & ego vision Rita Cucchiara ECCV W VISART 2014, Italy Gglass, Vuzix M100,Golden-i, Mobox+OdroidXU, MEG4.0 CV CHALLENGES IN EGO-VISION(2/3) 3. Recognizing head motion • • • • Head/body motion for outdoor summarization [ Peleg CVPR2014] Motion for indoor summarization [Grauman CVPR2013] Motion for supporting attention [Matsuo, CVPRW2014] Motion for SLAM as in robotics [Bahera, ACCV2012] 4. Recognizing objects • Objects useful for humans [Fathi, CVPR2013, Fathi, CVPR2011] • Objects in the hand [Fathi, CVPR2011] • Target tagging in the scene [Pirsiavash, Ramanan CVPR 2012] or Artworks in a museum…. Rita Cucchiara ECCV W VISART 2014, Italy CV CHALLENGES IN EGO-VISION(3/3) 5. Recognizing actions • Self-actions gestures [Kitani, CVPR 2013; Baraldi, EVW2014] • Actions of people, social actions [Ryoo, CVPR 2013; Alletto EFPVW 2014] • Actions in the environment (sport..) [Kitani, IEEE PC Magazine 2012] 6. Tracking: recognizing among the time • Tracking target objects • tracking face and people [Alletto, ICPR 2014] • Multiple target tracking Rita Cucchiara ECCV W VISART 2014, Italy EGO-VISION INTERACTION WITH CH Positioning: Recognizing targets indoor or outdoor HCI: Gesture recognition from egocentric video Recogition of emotions and feelings Experience augmentation: Recognizing visual/audio queries and interaction Rita Cucchiara ECCV W VISART 2014, Italy A VIDEO Video Gesture L. Baraldi, F. Paci, G. Serra, L. Benini, R. Cucchiara Gesture Recognition using Wearable Vision Sensors to Enhance Visitors’ Museum Experiences IEEE Sensors 2015 Rita Cucchiara ECCV W VISART 2014, Italy HAND SEGMENTATION IN EGO-VISION In ego-vision • Many luminance variations • Correcting strong camera/head motion • Recognizing ego-gesture from very few example It is an old problem, many approaches in different contexts: Skin classification: [Khan et al ICIP2010] Random forest: ( better than BN, MP,NB, AdaB…) Background subtraction after image registration [Fathi ICCV 2011] (assuming static bckg, hands with objects etc..) Generic object recognition : [Li, Kitani CVPR 2013] sparse feature selection and a battery of RF trained with different luminance conditions Rita Cucchiara ECCV W VISART 2014, Italy EGO-GESTURE RECOGNITION 1) (Ego-)Hand detection 2) (Ego-)Camera motion suppression 3) Feature extraction 4) Classification Rita Cucchiara ECCV W VISART 2014, Italy L.Baraldi, F. Paci, G. Serra, L. Benini, R. Cucchiara,"Gesture Recognition in Ego-Centric Videos using Dense Trajectories and Hand Segmentation« in Proc. of 10th IEEE Embedded Vision Workshop@CVPR2014 AN EGO-VISION SOLUTION Superpixel segmentation Superpixel descriptors Classification by Collection of RFs Temporal coherence Spatial coherence - SLIC (Simple linear Iterative clustering*) - K means in 5D (Lab+xy) *[Radhakrishna Achanta, Appu Shaji, Kevin Smith, Aurelien Lucchi, Pascal Fua, and Sabine Süsstrunk, SLIC Superpixels Compared to State-of-the-art Superpixel Methods, IEEE TPAMI, vol. 34, num. 11, p. 2274 - 2282, May 2012 Rita Cucchiara ECCV W VISART 2014, Italy [Serra ACM MMw 2013] [Baraldi CVPRW2014] AN EGOVISION SOLUTION Superpixel segmentation Superpixel descriptors - Classification by Collection of RFs Descriptors -mean and covariance in RGB LabH and HSVH 27 Gabor filters (9 orientation, 3 scales 7x7,13x13,19x19) - HoG Rita Cucchiara ECCV W VISART 2014, Italy Temporal coherence Spatial coherence AN EGOVISION SOLUTION Superpixel segmentation Superpixel descriptors Classification by Collection of RFs Temporal coherence Spatial coherence Classifier - Collection of Random forests - Indexed by a 32 bin RGBH - It encodes the appearance of the scene and the global luminance - Hp: bkg and hands changes colors accordingly Rita Cucchiara ECCV W VISART 2014, Italy Feature vector Scene luminance Global Luminance feature AN EGOVISION SOLUTION Superpixel segmentation Superpixel descriptors Classification by Collection of RFs Temporal coherence Spatial coherence Estimated priors Temporal smoothing in a window of k frames Posterior probability to be or not a hand pixel in a previous window Rita Cucchiara ECCV W VISART 2014, Italy AN EGOVISION SOLUTION Superpixel segmentation Superpixel descriptors Classification by Collection of RFs Spatial consistency Eliminate spurious superpixels Close holes Use grabcut using posteriori as a seed point Rita Cucchiara ECCV W VISART 2014, Italy Temporal coherence Spatial coherence HAND SEGMENTATION CONSISTENCY • Hand segmentation without temporal and spatial consistency • Hand segmentation with temporal and spatial consistency Rita Cucchiara ECCV W VISART 2014, Italy HAND SEGMENTATION Results shows that there is a significant improvement in performance when all the three consistency aspect are used together: • illumination invariance (II) • temporal smoothing (TS) • spatial consistency (SC). The method proposed by Li et al. is the most similar to our approach, nevertheless exploiting temporal and spatial coherence we are able to outperform their results. Rita Cucchiara ECCV W VISART 2014, Italy CAMERA MOTION SUPPRESSION 1) (Ego-)Hand detection 2) (Ego-)Camera motion suppression 3) Feature extraction 4) Classification Rita Cucchiara ECCV W VISART 2014, Italy CAMERA MOTION N • Camera (head) motion removal. • hands movement is usually not consistent with camera motion, resulting in wrong matches between the two frames. • For this reason we introduce a segmentation mask that disregards feature matches belonging to hands Dense motion Object detection Extract dense keypoints - Original frame sequence Output frame sequence without camera motion Estimate Homography Extract dense keypoints Rita Cucchiara ECCV W VISART 2014, Italy - Apply to EGO-GESTURE FEATURE • dense trajectories, HOG, HOG, MBH* extracted around hand regions. • Feature points are sampled inside and around the user’s hands and tracked during the gesture; inside a spatio-temporal volume aligned with each trajectory. • descriptors in a Bag of Words and then classified using a linear SVM classifier. *[Wang et al cvpr2013] Rita Cucchiara ECCV W VISART 2014, Italy DENSE TRAJECTORIES From * dense points ( but as in Shi Tomasi in ‘94 only if the eigenvalues of the autocorrelation matrix are every small) Points are connected in trajectory and normalized trajectory shape is used [Dalal and Triggs 2005 ] HOG static appearance [Lampert 2008] HOF with 9 bins: motion [Dalal 2004] Motion boundary histograms [*] H. Wang, A. Klaser, C. Schmid, and C.-L. Liu. Action Recognition by Dense Trajectories. In Proc. of CVPR, 2011 and IJCV2013. Rita Cucchiara ECCV W VISART 2014, Italy TRAJECTORY DESCRIPTION • Having removed camera motion between two adjacent frames, trajectories can be extracted. • Feature points are densely sampled at several spatial scales and tracked. Trajectories are restricted to lie inside and around the user’s hands. Without hand segmentation Rita Cucchiara ECCV W VISART 2014, Italy With hand segmentation TRAJECTORY DESCRIPTION • The spatio-temporal volume aligned with each trajectory is considered, and Trajectory descriptor, HOG, HOF and MBH are computed around it. • Since the histograms tend to be sparse, they are power-normalized to unsparsify the representation, while still allowing for linear classification. The function: is applied to each bin. • The final descriptor is the concatenation of the four power-normalized histograms. Gestures are eventually recognized using a linear SVM 1-vs-1 classifier. TD descriptors BoW HOG descriptors BoW HOF descriptors BoW MBH descriptors BoW Rita Cucchiara ECCV W VISART 2014, Italy Power-normalization and concatenation SVM 1-vs-1 EXPERIMENTAL RESULTS Datasets: • The Cambridge-Gesture database, with 900 sequences of nine hand gesture types under different illumination conditions; • Our Interactive Museum Dataset, an ego-centric gesture recognition dataset with 700 sequences from seven gesture classes performed by five subjects. • The EDSH dataset, which consists of three egocentric videos with indoor and outdoor scenes and large variations of illumination. Rita Cucchiara ECCV W VISART 2014, Italy GESTURE RECOGNITION Cambridge Gesture DB Results on the Interactive Museum Dataset using only 2 samples per class for training. [2] T.-K. Kim and R. Cipolla. Canonical correlation analysis of video volume tensors for action categorization and detection. Trans. PAMI, 2009 [3] Y. M. Lui, J. R. Beveridge, and M. Kirby. Action classification on product manifolds. In Proc. of CVPR, 2010 [4] Y. M. Lui and J. R. Beveridge. Tangent bundle for human action recognition. In In proc. of Automatic Face & Gesture Recognition and Workshops, 2011 [5] A. Sanin, C. Sanderson, M. T. Harandi, and B. C. Lovell. Spatio-temporal covariance descriptors for action and gesture recognition. In Proc. of Workshop on Applications of Computer Vision, 2013. Rita Cucchiara ECCV W VISART 2014, Italy LAST BUT… Rita Cucchiara ECCV W VISART 2014, Italy TRACKING: THE BIG CHALLENGE Rita Cucchiara ECCV W VISART 2014, Italy SINGLE TARGET TRACKING Tracking is the task of generating an inference about the motion of an object given a sequence of images *. in ego-vision is hard! * Arnold W. M. Smeulders, Dung M. Chu, Rita Cucchiara, Simone Calderara, Afshin Deghghan and, and Mubarak Shah, Visual Tracking: an Experimental Survey, IEEE TPAMI, July 2014. Rita Cucchiara ECCV W VISART 2014, Italy THE HARDNESS OF TRACKING Which is the invariance that can be perceived and maintained along the time? Tracking is hard as nothing is fixed: • Problems of lights: the target aspect, the illumination, • Problems of motion: the object/camera motion, • Problems of scene: the occlusion, the confusion... • …. Searching for the invariance in the video Rita Cucchiara ECCV W VISART 2014, Italy 14 TRACKING CHALLENGES IN 313 VIDEOS 01-LIGHT 02-SURFACECOVER 03-SPECULARITY 04-TRANSPARENCY 05-SHAPE 06-MOTIONSMOOTHNESS 07-MOTIONCOHERENCE 08-CLUTTER 09-CONFUSION 10-LOWCONTRAST 11-OCCLUSION 12-MOVINGCAMERA 13-ZOOMINGCAMERA 14-LONGDURATION Rita Cucchiara ECCV W VISART 2014, Italy EXPERIMENTAL RESULTS ON ALOV++ The upper bound, taking the best of all trackers at each frame 10% About the 30%, correctly tracked only [TST] A [FBT] [STR] [L1O] B [NCC] [TLD] C D E The lower bound, what all trackers can do 7% Rita Cucchiara ECCV W VISART 2014, Italy Survival curves by Kaplan-Meier See PAMI 2014 TRACKING IN EGO-VISION In egovision? • all the previous problems!! • relative motion • No motion of observer but motion of target • Motion of observer but fixed target • Motion of both observer and target • the dataset @Imagelab • EGO_GROUP • EGO_TRACK Rita Cucchiara ECCV W VISART 2014, Italy EGOVISION FROM A MOVING HEAD Rita Cucchiara ECCV W VISART 2014, Italy 0.40 TRACKING IN EGOVISION: EVALUATION Tracking results in the second scenario V2.2: tracking of a environmental point of interest. Target stays still but gets occluded and exits the camera FoV Color based trackers (HBT, NN) performs poorly due to the difficulty in discriminating the object based on color. Rita Cucchiara ECCV W VISART 2014, Italy TRACKING IN EGOVISION: EVALUATION Tracking results: table shows the F-measure for each video and each tracke a lot of work to do…r Scenari Still camera, o still person Video NN HBT TLD STR NCC FRT V1.1 V1.2 Moving camera, still person Moving camera, moving person V2.1 V3.1 V2.2 V2.3 V3.2 V3.3 0.5204 0.2793 0.2314 0.0472 0.1211 0.2552 0.0867 0.1565 0.5187 0.1177 0.0206 0.1602 0.0333 0.5786 0.1457 0.0973 0.4838 0.1767 0.5091 0.6372 0.4342 0.2446 0.0237 0.1303 0.6406 0.2397 0.0698 0.5745 0.0801 0.5532 0.0294 0.0879 0.4326 0.2251 0.4575 0.3769 0.0147 0.3607 0.1834 0.1118 0.2271 0.2138 0.1406 0.0294 0.0389 0.0984 0.1492 0.0756 Rita Cucchiara ECCV W VISART 2014, Italy TRACKING AND INTERACTING DICET Project: 2013-2015 tracking people and target for social interaction analysis On-line interaction with art Rita Cucchiara ECCV W VISART 2014, Italy CONCLUSIONS AND OPEN PROBLEMS • A Few conclusions: • Computer vision for multi-digitalization and interaction • • along successefull story 2D documents, 3D objects • Computer vision for real-time interaction • • • No hardware but software problem A long way ahead Interaction by egocentric vision is not an utopia anymore Rita Cucchiara ECCV W VISART 2014, Italy THANKS TO http://imagelab.ing.unimo.it Interdipartimental Research Center in ICT Tecnopolo di Modena Emila Romagna High Technology Network PEOPLE Rita Cucchiara Giuseppe Serra Marco Manfredi Costantino Grana Paolo Santinelli Francesco Solera Roberto Vezzani Martino Lombardi Simone Pistocchi Simone Calderara Michele Fornaciari Fabio Battilani Augusto Pieracci Dalia Coppi Patrizia Varini Stefano Alletto, Rita Cucchiara ECCV W VISART 2014, Italy