Application of Artificial Intelligence Technique in Bionic Eyes

DOI : 10.17577/IJERTCONV6IS14084

Download Full-Text PDF Cite this Publication

Text Only Version

Application of Artificial Intelligence Technique in Bionic Eyes

Application of Artificial Intelligence Technique in Bionic Eyes

Mohamed Mubeen. S

Department of Mechanical Engineering, Anjalai Ammal Mahalingam Engineering College,

Kovilvenni,Thiruvarur.

Abstract Biology can provide biomimetic components and new control principles for robotics. Developing a robot system equipped with bionic eyes is a difficult but exciting task. Researchers have been studying the control mechanisms of bionic eyes for many years and considerable models are available. In this paper, control model and its implementation on robots for bionic eyes are reviewed, which covers saccade, smooth pursuit, vergence, vestibule-ocular reflex (VOR), optokinetic reflex (OKR) and eye- head coordination. What is more, some problems and possible solutions in the field of bionic eyes are discussed and analyzed. This review paper can be used as a guide for researchers to identify potential research problems and solutions of the bionic eyes motion control.

  1. INTRODUCTION

    Eyes are the most important sensors for human beings during the process of information acquisition. More than 80 percent of information is acquired by eyes. The human visual system is highly developed and perfect after millions of years of evolution. Visual system can lock the object at the center of the retina (foveal area) even when the position of head or object changes drastically. This is of great significance for robots who always work in the bumpy and unstructured environment. Studying bionic eyes which can act like human beings is a difficult but exciting task. A robot system equipped with bionic eyes has a better performance in terms of perception and control.

    Main motion forms of bionic eyes include saccade, smooth pursuit, vergence, vestibule-ocular reflex (VOR), and optokinetic reflex (OKR) and eye-head coordination. Saccade is used to move eyes voluntarily from one point to another by rapid jumping, while smooth pursuit can be applied to track moving targets. VOR acts to stabilize retinal images by generating a compensatory eye motion during head turns. OKR can stabilize retinal images for gazing at rapid moving objects by nystagmus. OKR is driven by retinal slip while VOR is driven by head velocity signal. Commonly, two or more forms of motion work simultaneously. Besides, the binocular coordination and eye-head coordination are of high importance to realize object tracking and gaze control.

    The control mechanisms of bionic eyes have been studied for many years. Besides, some robot systems equipped with bionic eyes have been designed to implement these control models. Some typical systems of them are listed as following: The iCub robot [1] has 3 DOFs on the neck and 3 DOFs for the eyes. The KOBIAN robot [2] has 3 DOFs on the eyes, and 4 DOFs on the neck. The ARMAR robot [3] is designed to study gaze control. The Romeo [4] is a humanoid robot which employs 4 DOFs in the eyes. There are also some robot head systems such as the BARTHOC head[5] and the Flobi head[6] et al. Implementations of bionic eyes motion control mechanisms on these robot systems have validated their effectiveness for improving perception performance. However, restricted to the development of neuroscience, there are difficulties when imitating the performance of human eyes.

    The purpose of studying bionics is to imitate eye movements of primates to get a better performance in many aspects, such as gaze shifts and image stabilization. Most previous overviews about bionic eyes focused on the study of neurophysiology and designed control models to imitate these behaviors, but implementations of these control models on robot systems are always ignored. So, besides the current state-of-the-art of bionic eyes control mechanisms, their implementations on robot system are also discussed and analyzed in this paper. This review paper can be used as a guide for researchers to identify potential benefits and limitations of the bionic eyes study.

    The paper is organized as follows: In Section II, research status of different forms of bionic eyes motion is summarized, including saccade, smooth pursuit, VOR, OKR and eye-head coordination. Various models and their features are highlighted. What is more, implementations of control models on robots are presented. In Section III, some problems and possible solutions in the field of bionic eyes are discussed and analyzed. Finally, conclusions are drawn in Section IV.

  2. OVERVIEW OF MOTION CONTROL ON BIONIC EYES

    Biology can provide biomimetic components and new control principles for robotics. The motion forms of primate eyes include saccade, smooth pursuit, vergence, VOR, OKR and eye-head coordination. Studying on these motion forms can help researchers to improve the performance of the robots visual systems, including image stabilization, object tracking, navigation, and so on. Thanks to the efforts of many researchers, considerable control models for bionic eyes and their implementations are available.

    1. Models of Saccade

      Saccade is used to move eyes voluntarily from one point to another by rapid jumping. It is of great significance for robots to change their fixation point quickly. In the control models, saccade control system should act as a position servo controller to change and keep the target at the center of the retina with minimum time consuming.

      Young and Stark [7] proposed the sampled data model for saccade in 1963, which is shown in Fig.1. The circuit contains the dead zone and the INHBT (a device to inhibit the timing circuit), when the error exceeds a certain threshold, the pulse generator is triggered which causes a sample to be taken, at the same time, the INHBT element blocks dead zone for 0.2 seconds. The proposed sampled data model describes saccade by using discrete rather than continuous control loops. This model didnt take into account the actual brain structures, because so little was known at that time about the brain stem organization of eye movements. To solve this problem, Robinson [8] modified Youngs model by

      adding the premotor circuitry, which consists of medial longitudinal fasciculus (MLF) and neural integrator (NI). As shown in Fig.2, the MLF forms a lead network to compensate the lag of the plant, while the NI produces the position signal. Then the sum of MLF and NI pathway can be regarded as the input of the eyeball mechanics. The modifications are consistent with neurophysiology because of the fact that brain processes data in parallel.

      Pulse generator

      Dead 0.2s

      zone INHBT

      Target

      n

      position + Sampler 1 0.2s 2

      delay 2 2

      2 n s

      (FEL) strategy. In 2013, Wang et al. [11] designed an active vision system which can imitate saccade and other eye movements. The saccadic movements are implemented with open-loop controller, which ensures faster saccade eye movements compared with closed-loop controller. In 2015, Antonelli et al. [12] realized saccadic movements on a robot head by using a model called recurrent architecture (RA). In this model, the cerebellum is regarded as an adaptive element to learn internal model while brainstem is regarded as a fixed-inverse model. The experimental results on robot show that this model is more accurate and less sensitive to the choice of the inverse model compared with the FEL model. These models or methods for saccade and their features are summarized in Table1.

      Table1. Models or methods for saccade and their features see text for symbols

      – error s s

      Time and authors

      <>Models or methods

      Features

      1963,

      Young and Stark

      Sampled data model

      Discrete control loops,

      no brain structures

      1973,

      Robinson

      MLF and NI model

      Based on the Youngs model

      1997,

      Bruske et al.

      The FEL model

      Fast and accurate,

      implemented on the robot

      1999,

      Quaia et al.

      SC and cerebellum

      model

      The SC and cerebellum

      produce the saccade

      2013,

      Wang et al.

      Open-loop controller

      Fast, implemented on the

      robot

      2015,

      Antonelli et al.

      Recurrent architecture

      More accurate and less sensitive, implemented

      on the robot

      n

      Eyeye

      positionpositi on

      Fig.1 Youngs model for saccade, see text for symbols.

      0.2s Pulse MLF delay generator

      T

      Dead 0.25s e

      Target

      zone INHBT 1

      position error 0.2s Sampler 20 1

      + 1

      (

      sT 1)( sTe2

      + – delay s +

      NI

      e1 1)

      Eye

      positio n

      Fig.2 Robinsons model for saccade, see text for symbols.

      Both clinical and experimental evidences indicate that the superior colliculus (SC) and the cerebellum are important to produce accurate saccadic movements, and Quaia et al. [9] proposed a new model for saccade based on this fact. As shown in Fig.3, the solid line represents excitatory signals, the dotted line represents inhibitory signals, OPNs represents omnipause neurons, and MLBNs represents medium lead burst neurons. The superior colliculus pathway is parallel with the cerebellum pathway. The superior colliculus pathway provides a directional drive of eye movement while the cerebellum pathway keeps track of the saccade toward the target. Sum of these two drives is passed on to the motoneurons (MNs) and determines the velocity of the eyes.

      Cortex

      SC Cerebellum

      Study on saccadic movements has the longest history in control of bionic eyes. Saccade should act as a position servo controller to change and keep the target at the center of the retina. This is of great significance for bionic eyes to shift gaze quickly. Implementations of saccade on robots have been studied in the last twenty years. According to the implemented approaches, control models are designed to imitate the functions of brain regions such as superior colliculus, cerebellum and brainstem.

    2. Models of Smooth Pursuit

      Smooth pursuit is performed to track an object moving smoothly. Robots can track moving targets better by imitating the characteristics of smooth pursuit. In the control models, smooth pursuit system should act as a velocity servo controller to rotate the eyes at the same angular rate as the target. Fig.4 illustrates a simple model of pursuit [13], the velocity of the targets image across the

      retina,

      is taken as the major stimulus to pursuit and the

      OPNs

      MLBNs

      error that the system tries to minimize; then is transformed into an eye velocity

      MNs

      Fig.3 Quaias model for saccade, see text for symbols.

      command,

      Target

      e

      velocity

      , to be sent to motoneurons.

      A 0.13s E 1

      Besides, researchers have been studying the implementations of

      saccade on robots in the last twenty years. In 1997, Bruske et al. [10] implemented saccadic control on a binocular vision system by using the feedback error learning

      + sT 1 delays sTe2 1 Eye velocity

      Fig.4 A simple model of pursuit, see text for symbols.

      In 1986, the study of Robinson et al. [13] showed that smooth pursuit is a continuous negative feedback system responding to the eyeball velocity, which is different from saccade. Many current smooth pursuit models are descendants

      of the original model of Robinson. Based on the Robinsons model, Brown [14] added Smith predictor to cope with delays in loop control. In 1987, Lisberger et al. [15] proposed a model for smooth pursuit. The input of this model is the motion signal of retina. Then the motion signal is used to produce the position signal, velocity signal and acceleration signal of retina. Those signals are weighted and sent to an integrator to produce the velocity command of eye movements. In 1989, Deno et al. [16] applied dynamical neural network, which helps to unify two apparently disparate models of smooth pursuit and clarify dynamical element organization, to the smooth pursuit system. The dynamical neural network can compensate delays from sensory input to motor response. In 1998, Lunghi et al. [17] introduced a neural adaptive predictor which is previously trained to accomplish smooth pursuit.

      This model can explain humans ability to compensate the 130ms physiological delays when human beings follow the external targets with their eyes.

      Besides, some smooth pursuit models have been validated on the systems of bionic eyes. In 2001, Shibata et al. [18] implemented smooth pursuit control on a humanoid head. The control model can be divided into three subsystems: an inverse model controller, a target velocity predictor and a feedback controller. The results on robot head show a rapid convergence when learning the target dynamics. In 2012, Song and Zhang [19] proposed a binocular control model, which is derived from neural pathway, for smooth pursuit and other eye movements. In their smooth pursuit

    3. Models of VOR and OKR

      VOR acts to stabilize the retinal images by generating a compensatory eye motion during head turns. OKR occurs when the target moves rapidly in front of the eyes and it is driven by the retinal slip. OKR comes in two stages, a faster (saccade) and a slower (smooth pursuit).

      In 1992, Gomi and Kawato [20] presented an adaptive feedback control and feedforward control model for VOR- OKR. This model is based on the feedback error learning (FEL) scheme. By this method, it is concluded that the learning is accomplished in cerebellum region. In 1998, Raymond and Lisberger [21] found that the brainstem also plays an important role in the learning process. In 2001, Shibata and Schaal [22] added a learning controller as an indirect pathway for the VOR-OKR system. As shown in Fig.5, the learning controller takes the desired velocity and the estimated desired position as the input, and outputs necessary feedforward torque. VOR system is defined as a feedforward open-loop controller with an inverse model. OKR system is defined as a compensatory negative feedback controller for the VOR. This model can acquire good VOR performance after about 10 seconds of learning. Furthermore, it could converge to excellent performance after about 30 to 40 seconds of learning.

      learning

      experiments, the maximum retinal error is less than 2.2 degrees, which can keep the target in the field of view accurately. In 2013, Wang et al. [11] implemented smooth pursuit with velocity closed- loop controller, which is based on the image information. Their

      Head velocity

      network

      Indirect

      pathway

      experiments showed that the controller can adjust the eye velocity according to the distance between the target and the image center. These models or methods for the smooth pursuit and their features are summarized in Table2.

      -1 + VOR + Oculomotor

      + pathway + plant Direct

      pathway Eye

      position

      Head

      +

      Time and authors

      Models or methods

      Features

      1986,

      Robinson et al.

      The velocity negative feedback

      model

      Original model for smooth pursuit

      1987,

      Lisberger et al.

      Weighted signals

      Solves the conflict between high gain and

      large delay

      1989,

      Denoet al.

      The dynamical

      neural network model

      Adaptive

      1990,

      Brown

      Smith predictor

      Based on Robinsons

      model

      1998,

      Lunghi et al.

      Neural adaptive predictor

      Explain the compensation of delays

      2001

      Shibata et al.

      Be divided three subsystems

      Rapid convergence

      2012,

      Song and Zhang

      Derived from neural

      pathways

      Integrated, implemented

      on the robot

      2013,

      Wang et al.

      Closed-loop controller

      Adjust the eye velocity

      Table2. Models or methods for the smooth pursuit and their features

      position

      OKR

      + pathway

      Realized approaches show that control models for smooth pursuit should use some predictive controllers. From the cybernetic point of view, how to solve the delay problem is of great significance. Besides, coordination of smooth pursuit and other eye movements are worthy of further investigation.

      Fig.5 The FEL model with learning network.

      According to the clinical evidence, VOR includes rotational VOR (rVOR) and translational VOR (tVOR). They are controlled by semicircular canals and otolith organs, respectively. In 2005, Merfeld et al. [23] and Ramat et al. [24] proposed internal model and simple filtering model for rVOR and tVOR by clinical experiments, respectively. Porrill et al. [25] proposed a decorrelation control model for the VOR- OKR in 2004. The model is to remove any correlation between motor command and the variable that codes sensory error. It is worth noting that this model doesnt need the motor error signal by proposing a recurrent architecture. In 2010, Franchi et al. [1] compared the FEL and decorrelation model, and concluded that main difference between them is the learning control strategy. In the FEL model, a learning network is important and VOR collaborates with OKR sharing the frequency bandwidth response. In the decorrelation model, the OKR component is not taken into account and the model mainly focuses on the learning of cerebellar.

      In 2007, Khojasteh and Galiana [26] found that there are humans vergence eye movements during the VOR in the dark. It was believed that the eyes move in a perfectly conjugate fashion during the VOR. But Khojasteh et al. observed a significant vergence movement that is modulated with head

      velocity during VOR. This observation suggests a vestibular contribution to vergence.

      Many VOR-OKR control models have been validated on robot systems equipped with bionic eyes. In 2010, Kwon et al. [27] implemented the VOR-OKR-based vision tracking system on the mobile robot. The VOR-OKR concept is realized by using the robots motion information from the artificial vestibular system (AVS) sensor cluster and the vision information from the cameras. In 2012, Song and Zhang [19] implemented VOR-OKR movements on their binocular robot head by deriving the control model from neural pathways and pre-existing binocular models. Experiments show that better image stabilization result can be obtained by combining the VOR-OKR movements and the smooth pursuit movements. In 2013, Wang et al. [11] imitated the VOR movements using PID controller. And the angular velocity is obtained by placing an inertia tracker at the rotational axis of the head. Experimental results show that VOR can shorten the time of reaching to the target compared with eyes movements without VOR, which is from 5 seconds to 2.75 seconds. These models or methods for the VOR-OKR and their features are summarized in Table3.

      Table3. Models or methods for the VOR-OKR and their features

      Time and authors

      Models or methods

      Features

      1992,

      Gomi and Kawato

      Feedback and

      feedforward control

      Based on FEL

      2001,

      Shibata and Schaal

      FEL model with learning network.

      Add a learning controller as an

      indirect pathway

      2004,

      Porrill et al.

      Decorrelation control model

      Minimize the correlation function

      2005,

      Merfeld et al. and Ramat et al.

      Internal model and simple filtering model

      Explain the rVOR and tVOR

      2010

      Kwon et al.

      Use the robot motion information and vision

      information

      Stable, robust, implemented on the

      robot

      2012

      Song and Zhang

      Combined with the smooth pursuit

      Can obtain better image stabilization

      result

      2013

      Wang et al.

      Use the inertia trackers

      Shorten the time of

      reaching to the target

      Image stabilization is a necessary ability for bionic eyes. Many models and methods of VOR-OKR are proposed to realize image stabilization. According to the above researches, the learning process plays an important role in VOR-OKR control models. Besides, better image stabilization can be obtained by combining VOR-OKR and other eye movements.

    4. Eye-head Coordination during the Gaze Shifts

    Human beings shift fixation by changing their heads or bodys pose when the gaze shifts amplitude is too large, because it is not enough to keep target at the center of the retina only by saccade. There are two problems about the eye-head coordination: What areas in the brain play an important role for the eye-head coordination? What new models can be proposed to explain the gaze shifts better?

    For the first question, Freedman and Sparks [28] found that the desired gaze displacement command can be decomposed into separate eye and head displacement signals by the superior colliculus. Wang et al. [29] proposed a model that the gaze displacement signal is decomposed into eye and

    head contribution within the cerebellum. According to the above researches, both superior colliculus and cerebellum contribute to the eye-head coordination.

    For the second question, the optimal control is an efficient method to explain eye-head coordination during the gaze shifts. Gaze shifts should be as fast as possible in order to increase the time that the image is stabilized on the retina. So, Enderle and Wolfe [30] proposed an optimal control method that minimizes the time to reach the target, i.e. the minimum time rule. Harris and Wolpert [31] suggested another optimality principle called minimum variance rule, which minimizes the variance of the eye position. Kardamakis and Moschovakis [32] proposed a new method to understand the eye-head gaze shifts with the help of optimal control theory. They found that the gaze shifts obey a simple physical principle, i.e. the minimum effort rule. In this model, the squared sum of the eye and the head torque signals integrated over the movement period are minimized in order to obtain the optimal control signal. In 2011, Saeb et al. [33] proposed another optimal control method for eye-head coordination. The method combines the incremental learning with the optimality principle. This model can realize local adaption mechanism by minimizing the defined cost function.

    Many researchers have validated the eye-head coordination models on the systems of bionic eyes. In 2006, Maini et al. [34] implemented bionic eye-head coordination on their robot head. The control model used in the experiment is based on the methods in [35]. Both the eye and the head motors share the gaze displacement feedback signal while the gaze control system is driven by the collicular displacement signal. In 2011, Kido et al. [36] built a 7 DOFs robot to study eye-head coordination by gazing the moving targets. The control model is based on the methods in [37]. Thi model combines the eye- prioritized method and a mechanism that the head tends to return to the central position. In 2014, Vannucci et al. [38] proposed an adaptive controller to realize eye-head coordination with prediction of a moving target. The control model is improved based on the methods in [39], which combine the growing neural gas and the motor babbling technique. These models or methods for the eye-head coordination and their features are summarized in Table4.

    Time and authors

    Models or methods

    Features

    1987

    Enderle and Wolfe

    Minimum time rule

    Minimize the time to reach the target

    1998

    Harris and Wolpert

    Minimum variance rule

    Realize local adaption

    2006

    Maini et al.

    Gaze control model based on paper [35]

    Low residual errors, no residual oscillations of

    the head

    2009

    Kardamakis and Moschovakis

    Minimum effort rule

    Combine optimal control and system modeling of neural processes

    2011,

    Saeb et al.

    Incremental learning with optimality

    Minimize the cost function

    2011

    Kido et al.

    Control model based on paper[37]

    Simple, easy to implement

    2014,

    Vannucci et al.

    Improved based on paper[39]

    Fast, accurate, implemented on the robots

    Table4. Models or methods for the eye-head coordination and their features

    According to the above researches, both superior colliculus and cerebellum contribute to the eye-head coordination. The optimal control is an efficient method to explain eye-head coordination during the gaze shifts. Researchers have proposed some optimal control models based on different principles. Besides, how to implement these optimal control models should be paid more attention.

  3. PROBLEMS AND POSSIBLE SOLUTIONS Researchers have been studying the control mechanisms of

    bionic eyes for many years. Thanks to the efforts of these

    researchers, considerable control models for bionic eyes and their implementations are available. However, there are still many difficulties and problems because of two reasons: firstly, the study of bionic eyes is restricted by the slow development of physiology and neuroscience. Because most current control models for bionic eyes are derived from neural pathways, but the neural mechanisms are too sophisticated to imitate perfectly. Secondly, there are common challenges in the robot field such as speed, accuracy and robustness. In this section, some problems and possible solutions for bionic eyes are discussed.

    1. Information Fusion of the Monocular Vision and Binocular Vision

      Depth and motion information can help developing qualified behavior for bionic eyes. The platforms of bionic eyes are always equipped with two cameras, so perception methods based on binocular vision and those based on monocular vision can be applied to reconstruct depth and motion information for bionic eyes, simultaneously. The perception accuracy of binocular vision relies on the extrinsic parameters between two cameras, while the monocular vision method relies on motion parameters. Information fusion of the monocular vision and binocular vision may be an efficient way to improve the accuracy of perception. To the best of our knowledge, this has not been attempted before.

    2. Problems of the Eye-head Coordination

      It is of high importance for bionic eyes to change fixation and keep the target on the fovea. The eye-head coordination is necessary to realize this goal. Many current models for eye-head coordination cant be applied to real robot systems directly because the compensatory mechanism of eyeballs is not clear completely. In 2011, Milighetti et al. [40] derived kinematic mathematical models for eye-head coordination, which is a general control algorithm without neural model. But the calculation of inverse kinematics equation is too complicated to be accomplished in acceptable time. How to realize eye-head coordination rapidly and accurately is still worthy of noting. Besides, combination of saccade and VOR during the eye-head coordination is also worthy of further investigation.

    3. Problems of the Binocular Coordination

      Almost all of primates have binocular systems and many robot systems are also equipped with binocular vision to tracking targets. But there are only a little researchers take into account the binocular coordination. Most control models regard binocular vision systems as two unrelated cameras and control them independently. The binocular coordination can bring additional information for the control of bionic eyes. For

      example, one of the cameras loses target at a certain time while the other camera still captures target at the same time, then the first camera can be redirected to the target easier with the information provided by the other camera. How to realize binocular coordination and use these additional information are worthy of consideration.

    4. Microsaccade

      When humans try to fix their gaze on a certain target, the eyeballs are not stationary completely. There are also eyes movements called microsaccades, which typically occur during prolonged visual fixation. They are small- amplitude and involuntary eye movements just like miniature versions of saccades. In the last decade, researchers have been studying microsaccades. Evidences indicate that microsaccades contain depth information and can highlight edge information. This may provide a new approach to bionic eyes to get a better performance in image processing. Until now, microsaccade is still a largely unsolved topic and this movement is worthy of further investigation.

    5. New Methods

    Advances in many interdisciplinary fields, including optimal control, intelligence control and machine learning, are of great significance for the research of bionic eyes. The optimal control theory is an essential way to solve eye-head coordination problems, which can always obtain a better performance compared with the traditional PID control methods. Neural networks and other machine learning methods can also be applied to train models of bionic eyes. It is worth noting that some researchers have been studying bionic compound eyes which imitate insect eyes [41-43]. Insect compound eyes have wider field of view and faster speed compared with human beings. This study may provide new approaches for the design of bionic eyes.

  4. CONCLUSIONS

Control mechanisms and implementations for bionic eyes have been studied for many years and considerable models are available. In this paper, control models and their implementations for bionic eyes are reviewed. The research status of motion forms is summarized. Implementations of control models are presented as well. What is more, this paper discusses some problems and possible solutions in the field of bionic eyes. This review paper can be used as a guide for researchers to identify potential research problems and solutions of the bionic eyes motion control.

REFERENCES

[1] X. Tang, D. Pan, X. Zhang, L. Xi and W. Zhang, A novel receiver employing a compound-eye lens and a frequency domain synchronization algorithm for multiple-input single-output visible light communication system, in Proceedings of 2015 14th International Conference on Optical Communications and Networks, pp. 1-3, 2015.

[2] D. Keum, D. S. Jeon, M. H. Kim and K. H. Jeong, Artificial compound eye inspired by imaging principle of Xenos peckii, in Proceedings of 2015 18th International Conference on Solid-State Sensors, Actuators and Microsystems, pp. 403-406, 2015.

Leave a Reply