Research Article
A Model for Multimodal Humanlike Perception based on Modular Hierarchical Symbolic Information Processing, Knowledge Integration, and Learning
@INPROCEEDINGS{10.4108/ICST.BIONETICS2007.2421, author={Rosemarie Velik}, title={A Model for Multimodal Humanlike Perception based on Modular Hierarchical Symbolic Information Processing, Knowledge Integration, and Learning}, proceedings={2nd International ICST Conference on Bio-Inspired Models of Network, Information, and Computing Systems}, proceedings_a={BIONETICS}, year={2008}, month={8}, keywords={Bionics Building Automation Humanlike Perception Knowledge-based Systems Learning Multisensory Integration Symbolic Information Processing}, doi={10.4108/ICST.BIONETICS2007.2421} }
- Rosemarie Velik
Year: 2008
A Model for Multimodal Humanlike Perception based on Modular Hierarchical Symbolic Information Processing, Knowledge Integration, and Learning
BIONETICS
ICST
DOI: 10.4108/ICST.BIONETICS2007.2421
Abstract
Automatic surveillance systems as well as autonomous robots are technical systems which would profit from the ability of humanlike perception for effective, efficient, and flexible operation. In this article, a model for humanlike perception is introduced based on hierarchical modular fusion of multi-sensory data, symbolic information processing, integration of knowledge and memory, and learning. The model is inspired by findings from neuroscience. Information from diverse sensors is transformed into symbolic representations and processed in parallel in a modular, hierarchical fashion. Higher-level symbolic information is gained by combination of lower-level symbols. Feedbacks from higher levels to lower levels are possible. Relations between symbols can be learned from examples. Stored knowledge influences the activation of symbols. The model and the underlying concepts are explained by means of a concrete example taken from building automation.