- Simple ideas feed research on multimedia and computer vision
I’m an interaction designer and an intelligent web applications developer. My research interests focus on machine learning, collective intelligence, rich internet applications, social networks analysis and the semantic web.
I’m a PhD student at University of Florence. My main research interests are focused on application of pattern recognition and computer vision specifically in the field of video-surveillance with PTZ cameras, local pose estimation and 2D/3D face pose estimation.
I’m currently a PhD student at University of Florence. My research interests are focused on application of pattern recognition and machine learning, computer vision specifically in the field of human activity recognition.
I’m working as assistant professor at the Dipartimento Sistemi e Informatica of the University of Florence. My research work is in the field of Computer Vision and Pattern Recognition, and I mostly work on automatic video analysis, annotation and semantic transcoding.
I’m a developer and an interaction designer. My work focus on natural interaction and multitouch surfaces, rich internet applications and the semantic web.
- Andrea Ferracani
InstaMood is a Social Network for sharing photos with a mood using facial expressions. The App can be accessed using Facebook and allows to send photo messages selecting the mood that you are going to express taking a selfie.
Once logged-in users can select a mood they want to reproduce in their photo choosing between a set of six emotions represented through icons. The selection of the emotions has been done on the basis of the studies of Paul Ekman who has identified 6 pan-cultural expressions (angry, fearful, disgusted, happy, sad, surprised). Finally, the user can open the device camera, take a selfie trying to reproduce the emotion and send it to his Facebook friends. In order to exclude photos where no face is present the app uses the tracking.js lib which performs face detection. Photo messages can be sent to Facebook friends through the App. Messages are threaded.
The idea and and the interaction Design are by Andrea Ferracani and Daniele Pezzatini (Media Integration and Communication Center – UniFi, IT). Graphic Design and Development are by by Mariem Haida, Irene Rosadi and Elena Sesoldi.
The app has been developed at the Media Integration and Communication Center of the University of Florence by Mariem Haida, Irene Rosadi and Elena Sesoldi under the supervision of Alberto Del Bimbo, Andrea Ferracani and Daniele Pezzatini (MICC) for the Cource of Multimedia Design and Production.
Imaging Novecento is a mobile application for Android which allows visitors of the Museo Novecento in Florence, IT to frame some of the artworks in the halls of the museum which are automatically recognized by the application.
For each work the app provides some notes on the artist and his work style. In addition the user can select his own image from the device gallery and apply the style of the work on his picture.
Within minutes the processing result is sent to the user email. The photos can then be shared on the main social networks.
The app has been selected and developed in the context of “INNOVecento – Novecento Museum Innovation Lab”, a competition notice proposed by the Municipality of Florence, in order to enhance and continue the innovative vocation of its museums. The notice intends to call businesses and professionals to propose ideas and solutions based on ICT. Accepted proposals can be tested in the next months for a specified period of time at the Museo Novecento in Florence.
We presented our proposal as researchers of the NEMECH New Media for Cultural Heritage. NEMECH is a Competence Centre of Cultural Heritage established by Regione Toscana and activated by MICC – University of Florence, headed by the director Alberto Del Bimbo.
smArt is a low-cost framework to quickly set up indoor exhibits featuring a smart navigation system for museums.
The framework is web-based and allows the design on a digital map of a sensorized museum environment and the dynamic and assisted deﬁnition of the multimedia materials and sensors associated to the artworks.
The knowledge-base uses semantic technologies and it is ex-ploited by museum visitors to get directions and to have multimedia insights in a natural way.
Indoor localisation and routing is provided taking advantage of active and passive sensors advertisements and user interactions. In this way we overcome the Global Positioning System (GPS) un-availability issue in indoor environments.
The system will be presented at ACM Multimedia 2015, Brisbane, Australia in the Demos Track.
The system started as a project developed by the students Riccardo Del Chiaro, Franco Yang, Maurizio Sanesi and Alberto Merciai as final work for the Bachelor of Art at the Faculty of Engineering of University of Florence under the supervision of prof. Alberto Del Bimbo, and his assistants Andrea Ferracani and Daniele Pezzatini.