Next: References
Up: Cooperative Distributed Vision
Previous: Learning Cooperative Behaviors for
This paper described an overview of our five years project on
Cooperative Distributed Vision.
After one year research, we have obtained various promising results:
- From a scientific viewpoint, we proposed a functional dependency model
to explore the mechanism of the integration of perception, action, and
communication. While this model is not well developed yet, we can get
some basic characteristics of vacuous and embodied Active Vision
Agents:
- The essence of action is the internal state transition of an AVA.
The world state transition caused by the physical action should be
modeled by the side-effect of the action.
- While in vacuous AVAs no communication can be realized
without the message exchange, the communication between embodied AVAs
is best characterized by multi-channel communication links formed by
versatile combinations of perception, action and message exchange
processes.
- By analyzing the linear dynamic system model used for controlling
the soccer robot, the following points become evident:
- The reciprocal process consisting of the action-driven
perception and the perception-driven state-change followed by the
action-selection are the essential scheme of the model.
- To perceive complicated dynamic situations, the internal state
should have a certain amount of memory.
To make the proposed model really meaningful, we should augment it by
introducing temporal features, i.e. design a dynamic model. - From a technological viewpoint, we have the following results:
- We developed the Multi Focus Camera for real time 3D range
sensing, whose prototype showed its practical utilities as a compact
wide-range real time range finder.
- We proposed the Appearance Sphere Camera as a model of active
imaging method for observation stations. Sophisticated camera
calibration methods have been developed to make an off-the-shelf video
camera work as the APS camera. Experimental results proved its
practical utilities in generating a wide panoramic image by an active
pan-tilt-zoom video camera.
- Using the APS camera, we are developing various
algorithms for real time object detection and tracking. Experimental
results have shown that they can be used in the real world. We plan to
employ parallel processors to increase the processing speed and
multiple observation stations to realize the cooperative object
detection and tracking in a wide spread area.
- We proposed a behavior recognition method based on the
event driven selective attention mechanism. While the current system
is not so sophisticated, versatile behavior recognition will be realized
by incorporating multiple observation stations.
- We developed cooperative soccer robots and showed that
they can learn some primitive cooperative behaviors.
Intensive works on hardware/software developments, design of real
time video processing/generation algorithms,
communication protocols for cooperation, and visual learning are
being conducted to set technological foundations of CDV systems.
We plan to organize a series of workshops annually to promote
the CDV project. In addition to scientific and technological idea
exchange and discussion, collaborative liaisons with industries will
be established to apply our research results to practical fields. All
these project activities are listed on the project homepage (URL:
http://vision.kuee.kyoto-u.ac.jp/CDVPRJ/).
Finally, I would like to express my sincere thanks to Miss Hiromi
Taguchi for her help to prepare figures in the paper.
Next: References
Up: Cooperative Distributed Vision
Previous: Learning Cooperative Behaviors for