• Documents
  • Authors
  • Tables
  • Log in
  • Sign up
  • MetaCart
  • DMCA
  • Donate

CiteSeerX logo

Advanced Search Include Citations
Advanced Search Include Citations

Towards reliable multimodal sensing in aware environments," presented at 2001 Workshop on Perceptive user interfaces (2001)

by S Stillman, I Essa
Add To MetaCart

Tools

Sorted by:
Results 1 - 10 of 17
Next 10 →

Audio-visual multi-person tracking and identification for smart environments

by Keni Bernardin, Rainer Stiefelhagen - In Proceedings of the 15th international Conference on Multimedia (Augsburg, Germany). MULTIMEDIA '07. ACM , 2007
"... This paper presents a novel system for the automatic and unobtrusive tracking and identification of multiple persons in an indoor environment. Information from several fixed cameras is fused in a particle filter framework to simultaneously track multiple occupants. A set of steerable fuzzycontrolled ..."
Abstract - Cited by 22 (1 self) - Add to MetaCart
This paper presents a novel system for the automatic and unobtrusive tracking and identification of multiple persons in an indoor environment. Information from several fixed cameras is fused in a particle filter framework to simultaneously track multiple occupants. A set of steerable fuzzycontrolled pan-tilt-zoom cameras serves to smoothly track persons of interest and opportunistically capture facial closeups for face identification. In parallel, speech segmentation, sound source localization and speaker identification are performed using several far-field microphones and arrays. The information coming asynchronously and sporadically from several sources, such as track updates and spatio-temporally localized visual and acoustic identification cues, is fused at higher level to gradually refine the global scene model and increase the system’s confidence in the set of recognized identities. The system has been trained on a small set of users ’ faces and/or voices and showed good performance in natural meeting scenarios at quickly acquiring their identities and complementing the ID information missing in single modalities.
(Show Context)

Citation Context

...cation cues in time or across modalities. Stillman et al. [14] show a system for tracking, face detection and recognition of multiple users using a combination of fixed and PTZ cameras. In later work =-=[15]-=-, they extend their approach with the inclusion of microphone arrays for source localization, and show a framework, using a 3-layer hierarchical model and occupancy grids, for the fusion of multimodal...

Tracking People with Networks of Heterogeneous Sensors

by Alex Brooks, Stefan Williams - IN PROCEEDINGS OF THE AUSTRALASIAN CONFERENCE ON ROBOTICS AND AUTOMATION , 2004
"... This paper describes the theory and implementation of a system of distributed sensors which work together to identify and track moving people using di#erent sensing modalities in real time. Algorithms for detecting people using cameras and laser scanners are presented. A Kalman ..."
Abstract - Cited by 11 (0 self) - Add to MetaCart
This paper describes the theory and implementation of a system of distributed sensors which work together to identify and track moving people using di#erent sensing modalities in real time. Algorithms for detecting people using cameras and laser scanners are presented. A Kalman

A Self-Referential Perceptual Inference Framework for Video Interpretation

by Christopher Town, David Sinclair - In Proceedings of the International Conference on Vision Systems , 2003
"... This paper presents an extensible architectural model for general content-based analysis and indexing of video data which can be customised for a given problem domain. Video interpretation is approached as a joint inference problem which can be solved through the use of modern machine learning and p ..."
Abstract - Cited by 9 (0 self) - Add to MetaCart
This paper presents an extensible architectural model for general content-based analysis and indexing of video data which can be customised for a given problem domain. Video interpretation is approached as a joint inference problem which can be solved through the use of modern machine learning and probabilistic inference techniques. An important aspect of the work concerns the use of a novel active knowledge representation methodology based on an ontological query language. This representation allows one to pose the problem of video analysis in terms of queries expressed in a visual language incorporating prior hierarchical knowledge of the syntactic and semantic structure of entities, relationships, and events of interest occurring in a video sequence. Perceptual inference then takes place within an ontological domain defined by the structure of the problem and the current goal set.
(Show Context)

Citation Context

...agrammatic overview of the world model maintained by the sentient computing system. Combining vision with other sensory modalities is a very promising research avenue in ubiquitous perceiving systems =-=[8,35,9]-=-. Computer vision methods cans64 C. Town and D. Sinclair Fig. 4. The world as perceived by (a) users and (b) the sentient computing system. provide multi-modal human-computer interfaces with transpare...

Auditory Evidence Grids

by Eric Martinson, Alan Schultz - Proceedings of the IEEE Int. Conf. on Intelligent Robots and Systems (IROS , 2006
"... Abstract – Sound source localization on a mobile robot can be a difficult task due to a variety of problems inherent to a real environment, including robot ego-noise, echoes, and the transient nature of ambient noise. As a result, source localization data are often very noisy and unreliable. In this ..."
Abstract - Cited by 7 (1 self) - Add to MetaCart
Abstract – Sound source localization on a mobile robot can be a difficult task due to a variety of problems inherent to a real environment, including robot ego-noise, echoes, and the transient nature of ambient noise. As a result, source localization data are often very noisy and unreliable. In this work, we overcome some of these problems by combining the localization evidence over a variety of robot poses using an evidence grid. The result is a representation that localizes the pertinent objects well over time, can be used to filter poor localization results, and may also be useful for global re-localization from sound localization results.
(Show Context)

Citation Context

...bout walls and other obstacles distributed about an environment. An evidence grid then combines this data from a set of disparate, separated sensors to create a map of an indoor or outdoor environment=-=[3]-=-. While this work will be using evidence grids, there are other existing approaches in robotics to mapping out some form of the acoustic landscape. Noise Mapping[4] uses sound pressure level measureme...

Integration of angle of arrival information for multimodal sensor network localization using semidefinite programming

by Pratik Biswas, Hamid Aghajan, Yinyu Ye - In Proceedings of 39th Asilomar Conference on Signals, Systems and Computers , 2005
"... The problem of position estimation in sensor networks using a combination of distance and angle information as well as pure angle information is discussed. For this purpose, a semidefinite programming relaxation based method that has been demonstrated on pure distance information is extended to solv ..."
Abstract - Cited by 6 (0 self) - Add to MetaCart
The problem of position estimation in sensor networks using a combination of distance and angle information as well as pure angle information is discussed. For this purpose, a semidefinite programming relaxation based method that has been demonstrated on pure distance information is extended to solve the problem. Practical considerations such as the effect of noise and computational effort are also addressed. In particular, a random constraint selection method to minimize the number of constraints in the problem formulation is described. The performance evaluation of the technique with regard to estimation accuracy and computation time is also presented by the means of extensive simulations.
(Show Context)

Citation Context

...cularly useful in a scenario where we use sensors that are also able to detect mutual angles, like in image sensor nodes. There is a move towards towards creating designs and applications of networks =-=[15, 17]-=-. So there is also a need to develop localization techniques for such large scale heterogenous sensor networks using multi-modal sensing technologies. The idea is to move towards localization techniqu...

Performance analysis for automated gait extraction and recofnition . . .

by Michela Goffredo, Imed Bouchrika, J. N. Carter, M. S. Nixon - MULTIMED TOOLS APPL , 2009
"... ..."
Abstract - Cited by 6 (1 self) - Add to MetaCart
Abstract not found

Combining Environmental Cues & Head Gestures to Interact with Wearable Devices

by M. Hanheide - in Proc. Int. Conf. on Multimodal Interfaces, 2005
"... As wearable sensors and computing hardware are becoming a reality, new and unorthodox approaches to seamless human-computer interaction can be explored. This paper presents the prototype of a wearable, head-mounted device for advanced human-machine interaction that integrates speech recognition and ..."
Abstract - Cited by 4 (2 self) - Add to MetaCart
As wearable sensors and computing hardware are becoming a reality, new and unorthodox approaches to seamless human-computer interaction can be explored. This paper presents the prototype of a wearable, head-mounted device for advanced human-machine interaction that integrates speech recognition and computer vision with head gesture analysis based on inertial sensor data. We will focus on the innovative idea of integrating visual and inertial data processing for interaction. Fusing head gestures with results from visual analysis of the environment provides rich vocabularies for human-machine communication because it renders the environment into an interface: if objects or items in the surroundings are being associated with system activities, head gestures can trigger commands if the corresponding object is being looked at. We will explain the algorithmic approaches applied in our prototype and present experiments that highlight its potential for assistive technology. Apart from pointing out a new direction for seamless interaction in general, our approach provides a new and easy to use interface for disabled and paralyzed users in particular.
(Show Context)

Citation Context

...t reliably integrate speech and image understanding for natural interaction [1, 8, 10, 17, 19]. Other far advanced projects on intuitive interaction deal with tangible interfaces [3, 28], smart rooms =-=[2, 5, 27, 30]-=-, as well as wearable intelligent devices where space is becoming the interface [4, 11, 13, 21, 25]. A closer look at the cited contributions reveals that there seem to be two major directions in inte...

Multi-sensory and multi-modal fusion for sentient computing

by Christopher Town - Int. J. Comput. Vis , 2007
"... Abstract. This paper presents an approach to multi-sensory and multi-modal fusion in which computer vision information obtained from calibrated cameras is integrated with a large-scale sentient computing system known as “SPIRIT”. The SPIRIT system employs an ultrasonic location infrastructure to tra ..."
Abstract - Cited by 4 (1 self) - Add to MetaCart
Abstract. This paper presents an approach to multi-sensory and multi-modal fusion in which computer vision information obtained from calibrated cameras is integrated with a large-scale sentient computing system known as “SPIRIT”. The SPIRIT system employs an ultrasonic location infrastructure to track people and devices in an office building and model their state. Vision techniques include background and object appearance modelling, face detection, segmentation, and tracking modules. Integration is achieved at the system level through the metaphor of shared perceptions, in the sense that the different modalities are guided by and provide updates to a shared world model. This model incorporates aspects of both the static (e.g. positions of office walls and doors) and the dynamic (e.g. location and appearance of devices and people) environment. Fusion and inference are performed by Bayesian networks that model the probabilistic dependencies and reli-abilities of different sources of information over time. It is shown that the fusion process significantly enhances the capabilities and robustness of both sensory modalities, thus enabling the system to maintain a richer and more accurate world model.
(Show Context)

Citation Context

...ther sensory modalities [37]. Most such systems rely on integration with sound in the audible range via microphone assemblies, which has proven particularly popular for videoconferencing applications =-=[32, 6]-=-. There are also systems that have integrated vision assa secondary modality to an existing system, for example [3], which deploys visual gait recognition as an additional identifying cue to a system ...

Probabilistic Integration of Sparse Audio-Visual Cues for Identity Tracking

by Keni Bernardin, Rainer Stiefelhagen, Alex Waibel - Proceeding of the ACM International Conference on Multimedia , 2008
"... In the context of smart environments, the ability to track and identify persons is a key factor, determining the scope and flexibility of analytical components or intelligent services that can be provided. While some amount of work has been done concerning the camera-based tracking of multiple users ..."
Abstract - Cited by 4 (1 self) - Add to MetaCart
In the context of smart environments, the ability to track and identify persons is a key factor, determining the scope and flexibility of analytical components or intelligent services that can be provided. While some amount of work has been done concerning the camera-based tracking of multiple users in a variety of scenarios, technologies for acoustic and visual identification, such as face or voice ID, are unfortunately still subjected to severe limitations when distantly placed sensors have to be used. Because of this, reliable cues for identification can be hard to obtain without user cooperation, especially when multiple users are involved. In this paper, we present a novel technique for the tracking and identification of multiple persons in a smart environment using distantly placed audio-visual sensors. The technique builds on the opportunistic integration of tracking as well as face and voice identification cues, gained from several cameras and microphones, whenever these cues can be captured with a sufficient degree of confidence. A probabilistic model is used to keep track of identified persons and update the belief in their identities whenever new observations can be made. The technique has been systematically evaluated on the CLEAR Interactive Seminar database, a large audio-visual corpus of realistic meeting scenarios captured in a variety of smart rooms.

A Review of Decision Support Systems in Telecare

by Tasos Falas, George Papadopoulos, Andreas Stafylopatis
"... This paper presents an overview of the state-of-the-art on decision support systems (DSS) in telecare. The main aspect examined is the use of smaller subsystems— components in an integrated DSS, with emphasis on two application areas: medical home unit monitoring and real-time prioritisation of the ..."
Abstract - Cited by 2 (0 self) - Add to MetaCart
This paper presents an overview of the state-of-the-art on decision support systems (DSS) in telecare. The main aspect examined is the use of smaller subsystems— components in an integrated DSS, with emphasis on two application areas: medical home unit monitoring and real-time prioritisation of the alerts generated by them, and drug interaction checking. The paper suggests the development of an integrated hybrid telecare DSS synthesizing most of the technologies reviewed. Implementation issues are also examined, with an emphasis on the international trend towards the development of platform-independent, component-based, distributed software. KEY WORDS: telecare; decision support systems; alert prioritisation.
(Show Context)

Citation Context

...Decision Support Systems in Telecare 351 exactly its purpose. Therefore, wearable devices and multimodal sensing should be combined with the concept of ambient intelligence for telecare applications. =-=(10,11)-=- Drug Interaction Checking An area with significant importance and direct application of expert systems is the check for possible adverse effects with the drugs that are described for a patient. These...

Powered by: Apache Solr
  • About CiteSeerX
  • Submit and Index Documents
  • Privacy Policy
  • Help
  • Data
  • Source
  • Contact Us

Developed at and hosted by The College of Information Sciences and Technology

© 2007-2019 The Pennsylvania State University