
reason for the many existing professions. The basic 
structure for handling different task domains may be 
the same to a large extent. However, environments, 
objects and subjects likely to be encountered as well 
as typical behaviors of subjects may vary widely. 
Within each task domain there are characteristic 
maneuvers to be expected; therefore, driving on 
highways, on city roads, on the country side or in the 
woods requires different types of attention control 
and subjects likely to be detected. 
Learning which ones of these subjects with which 
parameter sets are to be expected in which situations 
is what constitutes “experience in the field”. This 
experience allows recognizing snapshots as part of a 
process; on this basis expectations can be derived that 
allow a) focusing attention in feature extraction on 
special events (like occlusion or uncovering of 
features in certain regions of future images) or b) 
increased resolution in some region of the real world 
by gaze control for a bifocal system.  
Crucial situation-dependent decisions have to be 
made for transitions between mission phases where 
switching between behavioral capabilities for the 
maneuver is required. That is why representation of 
specific knowledge of “maneuvers” is important.  
6 CONCLUSIONS 
In view of the supposition that human drivers will 
expect from ‘autonomous driving’ at least coming 
close to their performance levels in the long run, the 
discrepancies between systems intended for first 
introduction until 2020 and the features needed in the 
future for this purpose have been discussed. A 
proposal for a “Bifocal active road vehicle Eye” that 
seems to be an efficient compromise between 
mechanical complexity and perceptual performance 
achievable has been reviewed and improved. 
‘BarvEye’ needs just one tele-camera instead of more 
than seventy mounted fix on the vehicle body to cover 
the same high-resolution field of view. With respect 
to hardware components needed, there is no 
insurmountable barrier any more for volume or price 
of such a system, as compared to the beginnings. The 
software development in a unified design for detailed 
perception of individuals with their specific habits 
and limits continues to be a demanding challenge 
probably needing decades to be solved. Learning 
capabilities on all three levels of knowledge (visual 
features, objects / subjects, and situations in task 
domains) require advanced vision systems as 
compared to those used in the actual introductory 
phase. 
REFERENCES 
Bayerl S.F.X., Wuensche H.-J., 2014. Detection and 
Tracking of Rural Crossroads Combining Vision and 
LiDAR Measurements. In Proc. IEEE Int'l Conf. on 
Intelligent Transportation Systems, 2014. 
Bertozzi M., Broggi A, Fascioli A., 2000. Vision-based 
intelligent vehicles: State of the art and perspectives. 
Robotics and Autonomous Systems 32, pp 1–16. 
Burt P., Wixson L., Salgian G., 1995. Electronically 
directed “focal” stereo. Proc., Fifth Internat. Conf. on 
Computer Vision, pp 94–101. 
Dickmanns E.D, 1987: 4-D-Dynamic Scene Analysis with 
Integral Spatio-Temporal Models. In: Bolles RC, Roth 
B. 1988. Robotics Research, MIT Press, Cambridge. 
Dickmanns E.D., Graefe V., 1988. a) Dynamic monocular 
machine vision. Machine Vision and Applications, 
Springer International, Vol. 1, pp 223-240. b) 
Applications of dynamic monocular machine vision. 
(ibid), pp 241–261. 
Dickmanns E.D., 2007. Dynamic Vision for Perception and 
Control of Motion. Springer-Verlag, (474 pp). 
 
Gehrig S., Eberli F., Meyer T., 2009. A Real-time Low-
Power Stereo Vision Engine Using Semi-Global 
Matching on an automotive compliant FPGA. ICVS. 
Hirschmueller H., 2011 (Sept.). Semi-Global Matching - 
Motivation, Developments and Applications. Photo-
grammetric Week, Stuttgart, Germany, pp. 173-184.  
IJVAS-1, 2002. Vision for ground vehicles: history and 
prospects. Int. Journal of Vehicle Autonomous Systems 
(IJVAS), Vol.1 No.1, pp 1-44.  
IV’00, 2000. Proc. Internat. Symp. on Intelligent Vehicles, 
Dearborn (MI), with six contributions to Expectation-
based, Multi-focal, Saccadic (EMS-) vision: 
  1. Gregor R. et al.: EMS-Vision: A Perceptual System for 
Autonomous Vehicles.  
  2.  Gregor R., Dickmanns E.D.: EMS-Vision: Mission 
Performance on Road Networks.  
  3.  Hofmann U.; Rieder A., Dickmanns, E.D.: EMS-
Vision: Applic. to ‘Hybrid Adaptive Cruise Control’. 
  4.  Luetzeler  M.,  Dickmanns E.D.: EMS-Vision: Recog-
nition of Intersections on Unmarked Road Networks. 
  5.  Pellkofer M., Dickmanns E.D.: EMS-Vision: Gaze 
Control in Autonomous Vehicles. 
  6. Siedersberger K.-H., Dickmanns E.D.: EMS-Vision: 
Enhanced Abilities for Locomotion. 
Matthies  L., 1992. Stereo vision for planetary rovers: 
Stochastic modeling to near realtime 
implementation.  IJCV,  vol. 8. 
Niebles J.C., Han B., Li Fei-Fei, 2010. Efficient Extraction 
of Human Motion Volumes by Tracking. IEEE 
Computer Vision and Pattern Recogn. (CVPR). 
Pellkofer M., Luetzeler M., Dickmanns E.D., 2001. 
Interaction of Perception and Gaze Control in 
Autonomous Vehicles. Proc. SPIE: Intelligent Robots 
and Computer Vision XX; Newton, USA, pp 1-12. 
Roland A., Shiman P., 2002. Strategic Computing: DARPA 
and the Quest for Machine Intelligence, 1983–1993. 
MIT Press. 
BarvEye-BifocalActiveGazeControlforAutonomousDriving
435