Outils pour utilisateurs

Outils du site


news:journee_labex_esante:resume_bremond

Différences

Ci-dessous, les différences entre deux révisions de la page.

Lien vers cette vue comparative

Les deux révisions précédentes Révision précédente
Prochaine révision
Révision précédente
news:journee_labex_esante:resume_bremond [2014/10/28 11:53]
tigli
news:journee_labex_esante:resume_bremond [2014/10/28 11:56] (Version actuelle)
tigli
Ligne 1: Ligne 1:
 ===== Title : Scene understanding for Activity Monitoring ===== ===== Title : Scene understanding for Activity Monitoring =====
 +
 +
 +Intervenant : [[http://​www-sop.inria.fr/​members/​Francois.Bremond/​|François Bremond]], Directeur de Recherche INRIA, responsable du projet STARS d'​Inria Sophia Antipolis - Méditerranée
 +
 Scene understanding is the process, often real time, of perceiving, analyzing and elaborating an interpretation of a 3D dynamic scene observed through a network of sensors (e.g. video cameras). This process consists mainly in matching signal information coming from sensors observing the scene with models which humans are using to understand the scene. Based on that, scene understanding is both adding and extracting semantic from the sensor data characterizing a scene. This scene can contain a number of physical objects of various types (e.g. people, vehicle) interacting with each others or with their environment (e.g. equipment) more or less structured. The scene can last few instants (e.g. the fall of a person) or few months (e.g. the depression of a person), can be limited to a laboratory slide observed through a microscope or go beyond the size of a city. Sensors include usually cameras (e.g. omni-directional,​ infrared), but also may include microphones and other sensors (e.g. optical cells, contact sensors, physiological sensors, radars, smoke detectors). Scene understanding is the process, often real time, of perceiving, analyzing and elaborating an interpretation of a 3D dynamic scene observed through a network of sensors (e.g. video cameras). This process consists mainly in matching signal information coming from sensors observing the scene with models which humans are using to understand the scene. Based on that, scene understanding is both adding and extracting semantic from the sensor data characterizing a scene. This scene can contain a number of physical objects of various types (e.g. people, vehicle) interacting with each others or with their environment (e.g. equipment) more or less structured. The scene can last few instants (e.g. the fall of a person) or few months (e.g. the depression of a person), can be limited to a laboratory slide observed through a microscope or go beyond the size of a city. Sensors include usually cameras (e.g. omni-directional,​ infrared), but also may include microphones and other sensors (e.g. optical cells, contact sensors, physiological sensors, radars, smoke detectors).
 Scene understanding is influenced by cognitive vision and it requires at least the melding of three areas: computer vision, cognition and software engineering. Scene understanding can achieve five levels of generic computer vision functionality of detection, localization,​ tracking, recognition and understanding. But scene understanding systems go beyond the detection of visual features such as corners, edges and moving regions to extract information related to the physical world which is meaningful for human operators. Its requirement is also to achieve more robust, resilient, adaptable computer vision functionalities by endowing them with a cognitive faculty: the ability to learn, adapt, weigh alternative solutions, and develop new strategies for analysis and interpretation. ​ Scene understanding is influenced by cognitive vision and it requires at least the melding of three areas: computer vision, cognition and software engineering. Scene understanding can achieve five levels of generic computer vision functionality of detection, localization,​ tracking, recognition and understanding. But scene understanding systems go beyond the detection of visual features such as corners, edges and moving regions to extract information related to the physical world which is meaningful for human operators. Its requirement is also to achieve more robust, resilient, adaptable computer vision functionalities by endowing them with a cognitive faculty: the ability to learn, adapt, weigh alternative solutions, and develop new strategies for analysis and interpretation. ​
Ligne 23: Ligne 27:
  
 [[http://​cmrr-nice.fr/​sweethome/​|CMRR Nice]] [[http://​cmrr-nice.fr/​sweethome/​|CMRR Nice]]
 +
 +
 +
 +|||
 +|||
 +
 +[[ https://​www.tigli.fr/​doku.php?​id=news:​journee_labex_esante&#​programme_de_la_demi-journee|Retour]]
 +
  
news/journee_labex_esante/resume_bremond.1414493583.txt.gz · Dernière modification: 2014/10/28 11:53 par tigli