Home
Transparent Neural Networks, an Implementation JUAN
Contents
1. Transparent Neural Networks an Implementation Master s Thesis in Applied Information Technology JUAN SEBASTIAN OLIER Department of Applied Information Technology Division of Intelligent Systems Design CHALMERS UNIVERSITY OF TECHNOLOGY Gothenburg Sweden 2012 Report No 2012 010 REPORT NO 2012 010 Transparent Neural Networks an Implementation JUAN SEBASTIAN OLIER Department of Applied Information Technology CHALMERS UNIVERSITY OF TECHNOLOGY Goteborg Sweden 2012 Transparent Neural Networks an Implementation Master Thesisin Applied Information Technology JUAN S OLIER O JUAN SEBASTIAN OLIER 2012 ISSN 1651 4769 Technical report no 2012 010 Department of Applied Information Technology Chalmers University of Technology SE 412 96 G teborg Sweden Telephone 46 0 31 772 1000 Abstract The present work is related to a research project denominated Transparent Neural Networks Stranneg rd 2012 this project aims to propose a model capable of higher cognitive functions such as deductive and inductive reasoning by means of transparent simple and interpretable structures and functionalities This simplicity includes interactive building rules based on the mani
2. A toy example of this is an association that represents the concept apple let s assume that this concept is formed just by biding together the concepts apple shape and green color It is highly probable that green color is associated to many concepts whereas apple shape may be associated only by the concept apple if all the apples were green so the concept apple shape gives much more information to this association than what green color does In other words the relevance of the apple shape concept is much higher to the association apple than the color green To solve this it is suggested that weights for the real activity must be included in the edges The meaning of these weights as mentioned would be the relevance of each input which if managed correctly may keep the transparency of the model intact and give it better capabilities Another issue in this case related to controlling the growth of the network is that stability depends largely on the association threshold and a low threshold normally leads to an uncontrolled growth but even worst the problem is extended as the definition of low value in general is different for any given situation To partially cope with this problem an incremental threshold is proposed The idea is that as more abstract the concepts are the more stable they should be This means that at the bottom levels the association could be cre
3. Holcombe A O 2009 The Binding Problem In E Bruce Goldstein Ed The Sage Encyclopedia of Perception Ekbia H 2010 Fifty years of research in artificial intelligence In Cronin B Ed Annual Review of Information Science and Technology Volume 44 Medford NJ Information Today American Society for Information Science and Technology pp 201 242 Overskeid Geir 2008 They Should Have Thought About the Consequences The Crisis of Cognitivism and a Second Chance for Behavior Analysis The Psychological Record 2008 Vol 58 issue 1 pp 131 151 60
4. mouse over state and selected state When clicking over a node this will get selected and while the mose pointer is over it its information will be shown po Figure 15 Information of the selected node displayed when the mose is over it The same way as the node is selected it can be removed restarted or its label can be changed this is done by right clicking on the selected node A Remove node Remove input group Restart node Set label Figure 16 Edit menu for a selected node 41 Also when the node is selected its relation to other node can be edited by clicking the node to be related this action will display a pop up menu that gives the options Add edge Add inhibitory edge Remove edge If there is no edge between the two nodes an edge in purple color will show the possible connection to be created in order to visualize it easier o w o oo I Add edge Add inhibitory edge Remove edge 0 34 0 34 Figure 17 Connections menue deployed on a node to be related to the currently selected node Settings Ctrl T The settings dialog allows changing the basic parameters for the creation of nodes as well as options for the input stream reading and the random input generation In the upper part of the dialog the parameters for the creation of new nodes can be edited These are the Association Threshold for the minimum reverberation needed on a node to b
5. Figure 4 Depicts an inhibitory connection from the node with id 1 to the one with id 2 Activities As mentioned before there are two kinds of activity that propagates in opposite directions The main activity is called the real activity and is the one that propagates from the sensors forward till the deepest level When propagating the real activity the activation function of all the nodes at a certain level are evaluated taking into account the activity from the nodes in the previous levels so the activity is propagated level by level The second kind of activity is the imaginary activity which propagates backwards in the opposite direction than the real activity That means that the imaginary activity starts at the deepest level and propagates back until the first one However the imaginary activity at the deepest level or actually at every node that has no outputs will copy its real activity as the imaginary activity Each of the activities has different meaning On one hand the real activity is the natural response of the network to a given input and it is also the activity taken into account for the creation of new associations or concepts When evaluating the meaning of the real activity for each given node the amount of activity it has is related to how much of the concepts it associates were present in the inputs that relation depends on the activation function of each node and the interpretation may vary somehow based on that functio
6. activation of a concept or sensor given certain activation at a deeper level To do this the second sequence 1 2 4 is presented to the network half of the times the other one in this case 15 times Then the probability of occurring of each is different and in principle one must be half of the other When the network is fed with partial information Sequence 1 2 then the prediction on future input can be seen in the imaginary activity of the other two sensors in this cases as shown in figure 24 the imaginary activity of the third sensor is 0 5 whereas the one at the fourth it is 0 27 This implies that the probabilities learnt by the network into the weights of imaginary activities are tending to the actual probability of appearance of the sequences learnt 0 65 Figure 24 Imaginary activity as probabilistic inference for two different sequences that are parially equal 48 Generalization Generalization is performed by the Gaussian growth to show how this works a simple example is shown In this example the network has only two sensors and the generalization is to be made over two different classes Thus at the beginning a Complete Gaussian Node is manually added receiving inputs from both the sensors and afterwards examples from the two classes are shown from a distribution as the one that appears in the first table 0 2 men gt 5 1 1 0 35 0 2 Figure 25 Generalization by means og gaussian nodes Class 1 Cl
7. activity reverberation depends on the decay parameter in the following way pt Ple i 1 e v dilt 1 Rix t where R x t is the decay parameter of the edge To depict the behavior of the activity reverberation in relation to time as the parameter R t changes is depicted in the figure 6 Reverberation activity over time 12 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 Time steps Figure 6 Reverberation activity over time for different values of the reverberation parameter R t denoted R 27 To illustrate the temporal relationships a simple example of a sequence is shown in the figure 7 First a sequence of three consecutive inputs is shown in the first three time steps which happens while the recording signal is active After at the fourth time step the recording signal is deactivated thus the sequence is considered to be over and the association is formed as a new node This new node has three inputs and each one of the corresponding edges possess different reverberation decay parameter the first has a parameter 3 as it takes three time steps from the activation of the corresponding node until the whole sequence is over The same way the second edge will have a decay parameter 2 and the last one 1 Time step 1 Time step 2 Time step 3 1 1 a 1 E 0 0 1 0 10 Time step 4 gt 1 gt gt Figure 7 Example depicting the learning of a simple sequence in four timesteps by means of an
8. and the abstract methods to be implemented for new types Attributes Id It is a unique number that identifies the node and also corresponds to the index in the principal list of nodes in the network 34 state It is a double field that contains the real state of the node lt is updated at every time step lastState Retains the real state at the previous time step and is updated whenever a value is set for the atribute state predictedState Is a double field that contains the imaginary state of the node also called predicted state It is updated at every time step by the predict function lastPredictedState Retains the imaginary state at the previous time step and is updated when a value for predictedState is set reverberation It is a parameter used for creation of new association it copies the real state as long as it s value is higher than the one stored in reverberation and the recording signal emotional Impact is active Once the recording signal gets to 0 the reverberation is set to 0 again lastReverberation Retains the value of the previous value of reverberation It is updated each time a value for reverberation is set count Keeps track of the number of time steps the reverberation stays in an active state value higher than zero It starts to count only after the real state is set to a lower value than the one stored in the reverberation It is set to zero when reverberation changes from zero t
9. are and the reasons behind them the main goals of the present thesis is to implement a toolbox that allows its users to experiment with this kind of networks and draw conclusions about their behavior and utility Bearing that in mind and the fact that at the moment of the implementation the TNN research project is just starting the main value of this thesis becomes to generate feedback for further development on the ground concepts of the TNN The implementation was based on unpublished manuscripts that contained the general ideas and concepts of the TNN model which were evolving as the work was carried out Therefore to design the toolbox was needed to face a lack of specifications and take only as a major objective the concepts on modularity and transparency That implied to design a tool that could be flexible enough to fit the constant changes in the theory while producing useful and fast feedback The focus on modularity in the design allowed making the development flexible while fulfilling the TNN s main character of being built with elements and interactions as understandable as possible At the same time other important factor that led the design was to make interaction and building as easy and accessible as possible Thus there was a special emphasis on creating a friendly and simple to use interface that gives enough information to the user and at the same time allows creating exploring and evaluating TNNs fast and easily As h
10. by means of a more freely development Nevertheless the main fuscous on research has been set on sensory motor development leaving aside higher functions as the ones cognitive models and others pretend such as inductive and deductive learning or concept formation and manipulation Epigenetic robotics is in fact a source of inspiration for the model related to this thesis and presented below as even when the ideas are mainly focused on sensory motor approaches they seem promising for applications at higher functions as the ones aimed here Transparent Neural Networks The Transparent Neural Networks TNN model proposed by Claes Stranneg rd Stranneg rd 2012 is being developed since 2011 at Chalmers University of Technology and the University of Gothenburg Until the writing of the present thesis it is a theoretical model which has been presented in different conferences at the mentioned institutions as well as at the Lund University and the SewCog The TNN project attempts to develop a model with problem solving abilities achieved by means of transparent structures meaning that they are as clear as possible for the user at any time Thus the goal is to achieve transparency not only when designing a solution to a given problem but especially when the system has performed any kind of learning As described before when it comes to modeling traditionally the symbolic approaches are mainly concerned with deductive reasoning whereas emergen
11. disadvantages and discrepancies among them Emergentist approaches The emergentist approaches state that any behavior appears as a consequence of basic changes happening at a very low level cognition and abstract representations are therefore seen just as a consequence of processes carried out in underlying structures The main examples of this perspective are the connectionist approaches in these behavior emerge from the connections among simple processing units that acquire knowledge through experience by adjusting strengths between connections or either creating or removing them McClelland Botvinick Noelle Plaut Rogers Seidenberg and Smith 2010 The emergentist approaches propose alternatives most of the times inspired by nature or at least by our understanding of it The idea arises from the observation of how every structure around us emerges as patterns created by the interaction of smaller and simpler processes Consequently cognition also seems to be an emergent result of the interaction of simple and more understandable units Based on that perspective and considering that the brain is based on the interconnection of neurons it is assumed that is possible to have an emergent behavior from the simulation of simple units that at least in principal might mimic the behavior of real neurons This whole idea is the foundation for many connectionist approaches and specially the artificial neural networks These lasts attempt
12. equivalent to the reasoning itself Biasing the conception of reasoning in that way is indeed such a strong assertion that may cause a tremendously narrow view on the problem In any case the problem on biasing the analysis and therefore the assessment of performance of a model may come from any perspective In a certain way probability or mathematical models in general can be seen as a description of thought Chater Tenenbaum and Yuille 2006 and that assertion might be useful for many problems however one should bear in mind the it will only be a description and not thought itself either its equivalent To conclude aside of the assertion on reasoning or intelligence it can be said that probabilistic models have the abilities to infer and generalize and that allow solving problems and creating some useful behavior that may be of interest for particular applications such as those found in data mining and machine learning Symbolic approaches The symbolic approaches assume that cognition can be modeled by manipulating symbols and relations among them by means of structures and rules in this group can be included approaches such as the logic and rule based systems These models are mainly used as representation systems and are capable of inferences and deductive learning However they are criticized for lacking the ability of inductive reasoning as they normally are based on structures or concepts designed by the programmer but not lea
13. for proper performance so that the scalability and creation of solutions for dynamic and complex problems is not yet feasible with the current model The imaginary activity allows generating inferences and predictions by means of probabilistic relations therefore it should be used in the learning rules as it represents big part of the knowledge in the network and may be of utility to infer the relevance of a given input Despite the information given by the imaginary activity its interpretability in some cases is still too ambiguous as it encloses different concepts in just one parameter thus if inference and prediction or expectancy are to be interpreted separately there should be a difference in the treatment of this parameter for each of them The introduction of outputs and manipulation of inputs in relation to the knowledge needs of the system and specific goals may be needed to achieve automatic learning That is interaction with the environment may be mandatory to reach the ambitious goals of this model 56 Future work Here some issues considered to be relevant for future development of the TNN are mentioned as well as possible hints for their solution Firstly with the model as implemented when an association is created it asses the same relevance to each of its inputs however it can be sound to argue that not in every case the concepts associated give the same amount of information about the concept represented
14. idea presents to main problems in practice First the amount of time steps that have to be elapsed in order to delete certain association is defined in a too ambiguous way A wrong definition of this time may lead to instability in this case because there will be forgetting of relevant information as concepts are not presented in a certain period and also may avoid the creation of deeper concepts Secondly the basic idea of this rule may be ignoring that some associations might be extremely relevant but at the same time very rare then the importance of an association is not taken into account here Therefore it is suggested that certain kind of relevance is assed to each association in order to control forgetting which at the same time may help in controlling growth of the whole network but without the risk of deleting curtail information because of an arbitrary definition of time thresholds for forgetting In relation to the problems evaluated and the way they are presented more emphasis on different kinds of inputs and evaluations are highly recommended as the model and its performance as it is now is strongly directed to be a symbolic model with all what that implies The model should be available to create representations beyond the restrictions and reach the interpretation the user wants without a previous forced definition Finally as mentioned in the previous subsection there is a strong need for interaction in order to deve
15. must have in order to be considered when looking for new associations maxAssociationSize Used in the interactive mode is an integer that represents the maximum number of nodes that can be associated by a single node 37 Abstract methods void protected abstract void lookForAssociations This is the function called when associations are created automatically the process should perform both the search for possible associations as well as the addition of the corresponding nodes void train double emotionallmpact This method is called at every time step while the emotional impact is higher than zero and must include all the trainings procedures at the network level It also must call the train function of each node which is done by calling the method train of the class Node Basic management functions There are several basic functions that allow handling the network those that are necessary to build a network are addNode Node node o This function receives as input an object of any subclass of Node and adds it to the nodes list of the network but also to different lists depending of the kind of node It assigns an id to each node added taken from the index in the nodes list removeNode int nodeld o Receives the id of the node to be removed and performs the removal taking care of all the lists at which the node is included and the updating of the ids of the rest of the nodes The function will remove all nodes th
16. the transparency as it is one of the main objectives and in fact this characteristic is completely useful and understandable for the problems evaluated However all of the evaluations have been performed in relation to basic controlled and constrained problems reason why one could ask whether the ability for induction is still feasible if the data stops being constrained This question arises as the symbolic structures are always manipulated by the definition of a concept represented in the inputs but induction has not been clearly achieved for inputs which meanings or behaviors are not clear for the user at the design stage Similarly the growth control turns out to be a crucial issue for the equilibrium of the network but it may get quite instable as the data becomes more complex That fact sets a big challenge for developing automatic learning but again it may be challenging even for problems that include an unknown behavior Thus coping with the control problem is an objective to be stressed in further development as the reality is that almost any real problem may include unknown behaviors and unpredictable complexity But it is also important to highlight that there is an apparent tradeoff between the stability the control and the transparency that is not easy to deal whit in the current model Not controlling leads to instability and therefore interpretability and transparency get spoiled When studying other models it is found th
17. to mimic the behavior of real neural nets by interconnecting units that share information through weighted signals and activation functions The artificial neural networks have been seen in different ways as a feasible model of cognition and many claim that they not only model the cognition but also simulate the actual underlying processes Sharkey 2009 Nevertheless many arguments against connectionism and artificial neural networks strongly highlight that in comparison to real neural networks the models are oversimplified after all the real biological process is not yet completely understood and it could happened that certain assumptions may lack foundations For instance some other proposals on connectionism claim that the connections in a network should not rely on weights This is the case of the HTM model proposed by Numenta They argue that though in real neural 7 networks synapses might present a phenomenon similar to the one represented by the weights their values tend to be random or volatile therefore it cannot be safely assumed that calculations in the brain actually rely on those weights Numenta Inc 2011 Nevertheless the artificial neural networks have been subject of research for many years some have perceived them as model of cognition and others maybe more successfully as an optimization tool In the case of simulating cognition a good performance has been found based on fitting the network s response to s
18. Imaginary state Prediction 1 0 1 0 _Real state Figure 2 Description of the parameters appearing in a general node the node id the imaginary state and the real state Edges All nodes that are not sensors must always be connected to at least other node and this connection is represented by edges The Edges possess different characteristics firstly they may belong to two principal kinds depending on their direction they can be either those that are going out of the node or those that go into it However they are added by pairs so for every node going out of a node there is one going into another but graphically they are represented by the same connection This is done in order to treat separately the two kinds of activity that are there in the network All edges have a weight that can change over time but whether they change or not and the meaning they have depend on the kind of network being used The Edges also have an activity reverberation which means that after some activity is transmitted from one node to another the activity in the edge connecting them does not disappear immediately instead it fades down slowly according to a parameter learnt by experience called the decay parameter The reverberation activity of edge i at time t is denoted by b t Equally a vector of reverberation activities is denoted by B t Levels 20 As edges possess directions the way the network grows and propagates information is af
19. active That means that contextual relationships and inferences are being carried out all the time On the other hand to achieve both inductive and deductive reasoning two kinds of activity are used One of the activities in the network is called the real activity It is used to generate associations that may represent temporal relationships or specific concepts It also allows achieving inductive learning through the creation of deeper associations from more basic concepts scaling up the abstraction of the concepts at each level of association The other kind of activity is the imaginary activity this takes as basis the real activity and performs inductive reasoning by means of inferring causality predicting activity in future or deducing previous activity that could have led to the present state of the network At the same time this activity is capable of inferring missing information in an input or deducing possible relations by using existing associations When it comes to the construction rules they can be related to the way the nodes are added which can be manually done by the user or by an automatic addition partially assisted by the user taking into account information states in the network The way the activities are spread through the network as well as the construction rules and the characteristics in the elements of the network are detailed in the following section The implementation Given the general concepts of what TNNs
20. al to give a meaning to the values the weights reach after training and actually the meaning of the activity in a particular neuron apart of those in the inputs or outputs is not clear for the programmer Transparency is a desirable characteristic for any model as it makes it easy to explain maintain modify and verify Thus this is why the TNN attempts to maintain the transparency as a crucial issue in all the building element and learning rules in the model This is every element in the network must have a meaning or it may be easily inferred by the programmer equally any parameter that is modified by learning is to represent a simple and easy to comprehend relation among elements An introductory description of TNN The TNNs are networks constructed in stages by means of a small set of construction rules The construction rules are related to the addition of nodes and connections each node is to represent a clear function and its connections are to be easy to understand relations this way a compositional semantics in the networks is to be ensured It is important to make clear that even when the ideas behind TNN are related to cognitive modeling and are partially inspired by biology the aim is not to model any real neural system but instead the only concern of TNN is problem solving Again the main goal of the model is to achieve a transparent model capable of both inductive and deductive reasoning This transparency is to be achi
21. are available as classes that extend the abstract class Node There are also special relationships for the nodes of classes Input Association and Complete Gaussian There are special relationships because firstly the Association and Complete Gaussian are the only kind of nodes that are added automatically reason why there are special processes to check for the need of new nodes and the functions to create them Similarly the Input nodes sensors need to be tracked in order to update the network correctly since these nodes are updated in a different way than the rest There is also a need to track them in order to create groups and handle them which makes easier to add and remove sets of sensors easily 33 BufferNode _Butertode WLLL ComplementNode WLLL Y gt Relation gt Inheritance Figure 9 Classes diagram of the networks and nodes classes Main classes description The network is built as a collection of nodes of different kinds with individual characteristics behaviors and even some special functions to be handled However all the nodes possess a similar basic structure to which particular features are added and therefore all of them belong to the same basic class The class Node The abstract class Node has methods and attributes that are shared by all the nodes as well as some abstract methods to be implemented in any case when creating new types Below are listed the shared attributes
22. ass 2 Input 1 Input 2 Input 1 Input 2 Mean 0 750 0 183 0 190 0 750 Std deviation 0 041 0 062 0 070 0 041 After training is performed with a hundred inputs for each class the result is a network with two nodes each representing one class with the following parameters 49 Class 1 Node 2 Class 2 Node 3 Input 1 Input 2 Input 1 Input 2 Mean 0 74 0 19 0 19 0 74 Std deviation 0 063 0 081 0 086 0 063 0 16 0 2 3 7 0 82 p 0 82 Figure 26 Generalization of two different classes by means of gaussian nodes and gaussian growth The same way the weights for the imaginary activities of the sensors reflect basically the same values as the means learnt at the nodes This shows that in this case the imaginary activity reflects the expected value of the input from the activation of a node In other words if for example node 3 were to have a real state of 1 0 the imaginary activity at the inputs would be 0 19 and 0 74 respectively depicting the expected value of the inputs for each class 50 Drawbacks Through these examples it is shown the basic idea behind the attempt of TNN to achieve both kinds of reasoning in the same model while keeping the transparency However these are achieved when inputs are shown in a very controlled manner which can only be accomplished if the problem is known enough by the user However in cases where the problem is not completely known setting the basic parameters can become a complicated tas
23. association node 28 The parameters of each edge can change over time through experience tending to be the average time that it takes for the sequence or association to finish since the first activation of the edge Gaussian growth and generalization Other construction rule that can be used in the interactive mode is known as the Gaussian growth This construction rules creates nodes of the class Complete Gaussian Node which learning allows them to build concepts by generalizing characteristics of a set of inputs The main idea of this construction rule is to create a new Gaussian node when a given input is far from the characteristics learnt by the existing ones Therefore a new class will be created from a set of sensors by means of new a new node To set the sensors to be grouped in these classes they must be selected by manually adding a Complete Gaussian Node fed by those sensors as the seed of the whole learning Therefore the Gaussian growth is only performed in the level 1 of the network fed directly by sensors and consists on creating a new node if the ones existing do nat fit the current input by the following measure Mi 1 1 05 ea all o Xa i 1 gt 1 29 Network working cycle Updating the network is the process carried out at every time step The following description depicts the implementation in General networks for both the manual and the interactive modes The overall process of upda
24. at have no input or associations with just one after the removal So when the function is called more than one node can be actually removed addEdge int originNode int destinationNode double imaginaryWeight double realWeight Boolean inhibitory This function takes as parameter first the origin and the destination nodes id It also receives the initial weights for imaginary activity at the origin node and real activity at the destination node Finally it receives a Boolean value that indicates whether the edge is inhibitory or not When an edge is created an Edge object is added to each of the nodes one in the outgoingEdges list of the origin node and other in the incomingEdges of the destination node remove edge int originNode int destinationNode o Given the ids of the origin node and the destination node the edge is removed if it exists This is done by removing the objects at the corresponding edges lists of each node 38 The Tool box user s manual In this section the main issues on the usage of the toolbox are explained The functionality of the toolbars and menus are shown as well as the characteristics of the interface and the way information is displayed The toolbar is divided in two smaller ones the first one is file toolbar that is the one with which the basic actions over files can be performed these actions are described below New network Open networkk Save network Open input file Add nodes Figure 10 The
25. at the problem on controlling growth has been addressed by many and they always end up facing the so called combinatorial complexity or the biding problem These problems arise when models create concepts by binding representations of intrinsic characteristics in the entities to be represented This idea becomes problematic as the representations include more and more characteristics since the possible number of combination increases exponentially Therefore in the field there has been a quest for reducing the complexity and the amount of associations created when these kinds of problems arise One important process with which possible solutions to the biding problem in real biological systems have been described is the need for attention Holcombe 2009 Attention can be described in many ways and the real process is not completely known however its possible need for solving the biding problem implies certain control on which inputs and the way they are bind at a given time This overall idea of attention may have direct relation to the control signals in the TNN model which allow proper performance by stating when to associate and selecting the inputs that are to be related In that sense the control and the 53 constraining of inputs proposed is arguable in terms of solving the binding problem but then again it is not a feasible solution for automating the process In order to achieve automatic selectivity for associations many other fe
26. ated and also deleted easily but the deeper the concepts being related are the more the threshold should be increased to ensure that associations created bind clearer concepts at each level However there will still exist a dependence on the nature of the problem being addressed It was also shown that by means of the imaginary activity it is possible to achieve inferences and predictions by means of probabilistic relations however these are related in the same way and represented by the same unique value This unique value affects the interpretability of the results as it is hard to tell what of the possible meaning the activity has at a given point or in fact the meaning may always arise from a mixture of all possible interpretations which is against the supreme goal of transparency Therefor a different treatment for either the interpretation or the computation of the imaginary activity is suggested 57 Something also related to the amount of associations created and its control is the forgetting rules that have to be implemented The goal of these rules is to delete certain associations that are not really relevant as may have been caused by noisy inputs or other situations However though not currently implemented the forgetting rules in the model have been proposed to be simply based on the usage frequency of the associations meaning that if an association does not get activated in a long time it will be deleted That
27. atures have to be included in the model Some other issues are also related to the biding problem beyond the complexity and growth of the structures Typical examples are connected to the ability of assessing proper meaning to the associations and are those including relational statements of the kind Mary loves John That relation could be seen as two subject or concepts bind by a relation called love or a relation among three particular concepts that are bind together however the original statement does not imply the complementary John loves Mary but when the relationship is created as described both the statements can mean the same which is not necessarily the case For this example the TNN model may manage the two possible statements by means of two different associations having the possibility of interpreting them differently This is possible if the statements are presented as different sequences then each sequence will represent a different concept to which a distinctive meaning could be assessed However this implies a symbolic manipulation that requires that the three concepts are clear and again the design is limited to a symbolic well understand behavior of the inputs In general this discussion and the development of the model are related to a broader set of questions on the need for the development of models capable of really creating and understanding concepts and not only perform some manipulations on specific
28. ay of getting to visualize the problems and possible behaviors Exploring the problem in this mode can help to find a reasonable starting point for the further growth based on an interactive construction Interactive mode In contrast to the Manual mode in the interactive mode the network can be modified automatically by adding new nodes and connections depending on the need However it is not completely automatic since is the user that controls when the network should look for new associations The user has the ability to set the network in a recording mode and stop it when needed In this way the network will look for associations presented in between the time the recording signal is active but this search in only performed in the time step at which the recording signal stops Associations in the interactive mode To better understand the idea with the interactive mode the definition of association has to be enhanced An association in the sense used in this particular mode is the formation of a node that represents a relationship between the activities of two or more nodes limited by a maximum that can be set These relationships may represent a simultaneous activation of nodes or a temporal relation among them though in general is the same behavior The temporal relations refer to the situations at which one node or a group of nodes get active or increase their activity after other has done the same This may include many steps and re
29. cept from two previous ones 45 When the network is fed again with just one of the two basic concepts one can see how the composed concept gets activated to a 50 This partial activation allows the network to predict possible associations that could appear by means of the imaginary activity as shown in figure 22 The imaginary activation at node 5 representing the second concept as well as in the two inactivated sensors represents an inference from the known association between the two basic concepts This activation indicates that there is a possibility of the two basic associations of appearing together which is an association that has been learnt and is used for inference through imaginary activity in this case O 4 6 _ 1 0 p 1 0 0 5 24 0 0 5 05 1 Fem O 1 a iS 0 0 ie Figure 22 Partial activation of a deep concept when only one of the concepts associated is presented in the input Temporal associations Similarly association will be created when the sensors are activated sequentially and in this case the maximum activation of the association will be reached as the sequence is completed after increasing after each time step This means that as more information in relation to the sequence learnt more real and imaginary activities there will be in the corresponding association However if all the elements of the sequence are presented in different order than the one learnt the association will incr
30. ction and concepts Anthony F Morse Joachim de Greeff Tony Belpeame and Angelo Cangelosi 2010 Evidently pre programing for specific behaviors cannot give solutions to scalability problems as the systems are expected to work in too complex and unpredictable environments that as the limitation set by constraints and assumptions made by the programmer usually fail when the systems are faced to real problems Therefore the best is to make the systems in charge of their development by giving them the ability of verifying their own learning and the possibility of growing their cognitive structures freely towards broader goals Stoytchev Alexander 2009 However as the systems are expected to develop by means of interacting with the environment a clear limitation and crucial aspect in the design is the actual body of the robot in the sense that it will constraint any interaction and therefore the whole process Body and brain cannot be separated and at the end is the body what shapes the brain Asada Hosoda Yasuo Hiroshi Toshio Yoshikawa Ogino and Yoshida 2009 reason why in epigenetic robotics the design of the body in terms of sensors and actuators plays a very important role in the abilities that can be achieved and therefore great deal of the research is focused on this fact 14 Epigenetic robotics is a relatively new field but it has shown interesting results as many limitation and constraints of other approaches are overcame
31. ctive in turn proposes that cognitive systems use specific symbols as a representation of knowledge and find solution by carrying out processes on these representations Complementarily the emergentist perspective proposes that the knowledge is represented in a distributed manner into basic elements and processing is carried on this distributed knowledge in a complex and meaningful way Troy D 2003 Another way of classifying the models is by the top down and bottom up differentiation Top down perspectives assume that the basis of cognition lays on the symbolic abstractions and therefore only that is needed to achieve intelligent behavior thus the relevance of whatever structure that is below the whole process can be neglected On the other hand bottom up perspectives assume that intelligence and cognition emerge from the behavior of atomic components in a structure and the way they relate to each other They argue that it is possible to achieve abstract associations from basic processes at the bottom of a structure McClelland Botvinick Noelle Plaut Rogers Seidenberg and Smith 2010 However these classifications are generalizations there are also models that combine the approaches in attempts to achieve better results but they do not fit completely in any on the groups above To better understand the characteristics of each of the approaches they will be described below by giving some notions about their utility as well as advantages
32. des boolean addincomingEdge int origin double weight boolean inhibitory This function must return a Boolean that indicates whether the edge was added or not There is a default function called createlncomingEdge that receives the same parameters and returns a Boolean The implementation of addIncomingEdge can be just a call to createlncomingEdge however it is left abstract in order to allow certain rules for each node for example at a buffer node it is not possible to add more than one edge then this function is used to add that rule however to add the corresponding object Edge createlncomingEdge must always be used restartFunction When a node is restarted it goes back to its initial state undoing any kind of learning and resetting default values When a node is restarted all the edges and principal attributes are restarted however if more parameters are included in certain kinds of nodes this function should include the restart procedure of those parameters if needed String getInfo It returns a string where some information about the node can be added This info is what will be displayed in the interface when checking the information of the node It has no relation to the actual functioning of the network 36 The class Network The same way as in Node the abstract class Network has methods and attributes that are shared by any possible kind of network It also has some abstract methods to be implemented w
33. e considered for new associations and the Maximum Association Size which determines the maximum number of nodes admitted per association when added automatically It also can be selected whether or not to Allow Gaussian growth for the creation of new Complete Gaussian Nodes 42 Association Threshold 0 5 Maximum Association Size 7 J Allow gaussian growth Figure 18 Association parameters in the Settings box The second part in the dialog allows changing the number of times the input file is read when the complete stream button or random streams are used lt can be also specified whether the randomly generated inputs are binary or not When not selected the random inputs generated will be numbers in the interval 0 1 otherwise they will be binary values 0 or 1 Number of iterations over input file 1 Generate random binary inputs Figure 19 Inputs reading and generation parameters in the settings box Zooming and exploration For zooming the zoom bar or the scroll wheel of the mouse can be used which will enlarge the size of the nodes and therefore the whole network To explore the network this can be moved throughout the scree by clicking at any empty space and moving the mouse while still clicking the network will follow the movement of the mouse 43 Results The objectives of the TNN model as stated in the description are mainly related to both the transparency and the ability to perform deductive and i
34. ease its real activity but will never 46 reach the same value as in the case of the sequence being presented as it was learnt The sequence taught to the network in this example is simply three sensors being activated consecutively In figure 23 after training is done one can see how the activation in the association node increases accordingly to the amount of information as the sequence learnt is shown again This increase appears both in the real and the imaginary activities showing how through imaginary activity prediction on future and inference on past are performed Time step 1 Time step 2 e RC z o r E d d a e gt A _T z 1 0 de 0 65 0 0 65 0 65 Time step 3 o D no Figure 23 Depicts the process of activation of a temporal association when the sequence it associates its shown 47 Predicting the most probable input form partial information In this example two different sequences are shown to the network Both the sequences are of a three time steps length and both include the first two sensors as the beginning of the sequence Then the only difference between them is the last element being in one case the third sensor and in the other the fourth one These examples can resemble the two number sequences 1 2 3 and 1 2 4 which only differ on one number but one can be more probable to occur than the other Here is shown how the imaginary activity also represents the probability of
35. ed while there is a positive emotional impact in the input To create associations the process is as follows The search is performed after an interval of time steps has elapsed During the interval an attribute of each node called reverberation of the node is set to the maximum real state the node reaches within the interval Similarly at the time step at which the real state is found to be lower than the reverberation the attribute called count starts to keep track of how many steps pass from that event until the end of the interval Once the interval is finished the associations search starts it begins at the deepest level going backwards to the sensors level but it stops wherever an association is created The search is carried out basically by grouping all the nodes that have first a reverberation value at the previous time step higher than 31 an association threshold given by the user and secondly if the predicted stated is lower than the last reverberation or the node has no outputs In principle the group can be of any size but only a maximum number is associated depending on a parameter of maximum _ association size that is determined also by the user The nodes of this final group are organized by the count parameter and so are added to a new association node This allows differentiating associations including the same nodes but different time order In case it is found that an association already exists an update is perform
36. ed on the reverberation parameters of the incoming edges in the node representing that association This update is performed taking into account the count parameters of the nodes feeding the association tending to the mean of all the examples seen Train When training the network two processes can be performed the principal one has to do with the weights training whereas the other is carried out at each node and is related to the update of certain parameters In the manual mode no training is performed on the weights meanwhile in the interactive modes the only weights changes are those related to the imaginary activity These are trained to represent a probabilistic relation between the activation of the nodes linked by the edge The idea is to get a parameter that encloses how probable is that a node a feeding node b was active during the recording interval if b had certain activity during the same interval The other kind of training occur for some nodes that have parameters to be adjusted to the inputs the Gaussian nodes are the only ones that train parameters as they adjust their mean and standard deviation values at each time step unless a Gaussian growth is performed 32 General description of the implementation The implementation of the toolbox is divided in three main block as depicted in the diagram below The principal block is the network in which the functionality and algorithms as well as the structure and management of
37. el 0 or sensors level is first updated by copying the values from the input From then on the nodes are updated by using their corresponding updating function r t level by level After a whole level has been updated the inhibitions of the nodes are carried out This is done at each level in order to avoid the propagation of activity of nodes that are to be inhibited Update imaginary states Predict Once the real activity in the nodes has been set the imaginary activity is propagated backwards from the last level to the sensors The calculation is done as was described in the imaginary activities section Look for associations When the recording signal is being used this function performs the search for new possible associations when certain behavior in the signal is met The recording signal The recording signal must be used to set the network in the interactive mode meaning that associations are to be searched and automatically created when the user decides to The recording is done for time intervals that are specified by this signal The interval starts at the time step in which the signal changes from 0 to any higher value in the interval 0 1 and it finishes when the signal goes back to 0 In the interactive mode this input is referred to as the recording signal however in the implementation and the interface of the toolbox this parameter is known as the emotional impact of the input In other words the recording is perform
38. ell that if an input is to last longer than other it simply has to be repeated several times in the input stream In other words if an input is repeated in several time steps this will keep the network in the same state after it is stable 25 Networks construction modes To build networks different construction rules can be used but which and how are used depend on the construction mode selected The two modes existing in the current implementation are the Manual and the Interactive modes Manual mode The most basic construction mode that can be used to create networks is the manual one The construction of a network in this mode depends completely on the user The architecture of the networks does not change while they are being used but only as the user decides to add or remove elements In this mode the user chooses to add any kind of node by connecting them through edges from whatever node that already exists unless the node added is an input This allows a complete and easy understanding of the network though at the same time its usage is limited to a rigid architecture It works to evaluate and visualize how the activities spread through a network s architecture but no automatic addition of elements is performed This mode is used mainly to propose anatomies and check their performance Building a solution might require a complete understanding of the problem reason why is not suitable for this purpose but instead it is a good w
39. els are defended as being capable of yielding great flexibility for exploring the representations and inductive biases that underlie human cognition Griffiths Chater Kemp Perfors and Tenenbaum 2010 That assertion is based on the assumption that whatever behavior a system displays its causes can be easily described by means of probabilities This flexibility at the time of exploring inductive behaviors is a characteristic that represents an advantage when it comes to fully understand the system and what it represents In fact this idea has been used against Bottom Up and some connectionist models by arguing that even when both kind of models could successfully address similar problems the way emergentist models solve them is not necessarily as understandable or transparent to the user as a probabilistic model could be Griffiths Chater Kemp Perfors and Tenenbaum 2010 However reality is that mathematics behind probabilistic inferences can easily go beyond unaided intuition and even simple rules can become intractable as models are scaled up to fit real world problems McClelland Botvinick Noelle Plaut Rogers Seidenberg and Smith 2010 That may contradict the claim of probabilistic approaches being capable to draw more understandable descriptions of reasoning and cognition Nonetheless this leads to a more general topic than the one concerned to this document but that still affects the fundamentals of the Transparent Neural Netw
40. ete Gaussian Node Association Node Number of nodes Node that sets as its own state the minimum value among all the inputs Figure 12 The add nodes dialog The second part in the division of the bar is the inputs toolbar which handles the way the inputs are fed into the network It has four buttons two of them are to read streams from the input files and the other two are to generate random inputs 0 ee Figure 13 The inputs tool bar in the tool box From left to right the buttons in this bar are Complete stream button This will read the whole input file feeding the network step by step It will do that several times depending on the parameter Number of iterations over input file that can be set by the user in the settings dialog Step button This will only read one line in the input file at the time so it goes one step at the time through the input stream 40 Random stream button This will generate a random stream whit a number of steps equal to the same parameter used for the Complete stream button Random step lt will generate a single random input While exploring the network and manipulating the nodes they will be shown in three different ways in relation to the mouse actions The three states are normal mouse over and node selected 0 56 0 56 0 56 0 56 0 56 0 56 Figure 14 Different states for interaction with the nodes in the tool box From left to right normal state
41. eved by the limitation in the construction rules that ensure the interpretability of every element Therefore the basic elements are to represent clear concepts and their association to others must be clear relationships The most basic elements of the networks are the nodes which in the model are to represent concepts learnt by experience This way each node in the network is aconcept and is related to other by means of connections called edges The relations between concepts by means of the edges and the information spread through them further called activity allow the formation of conceptual relationships that emerge contextual meaning for each node This permits that concepts with partial information are retrieved or that inferences of concepts contextually connected are made even when the explicit information that elicits them is not in a given input To illustrate this imagine a concept representing a physical object and therefore its activation is elicited by sensing the physical characteristics of the real object 16 when presented as an input However this concept could also be related to a concept that represents the name of the object in the form of a word which activity is elicited by the sound that corresponds to the word presented as an input Then even when the physical characteristics of the object are not present in the input activity in the concept representing it may be elicited if the concept representing the word is
42. existing concepts or sensor nodes The sensor nodes are the inputs to the network and are how the network is fed and receives information from environment Sensor node _ General node concept Figure 1 The two main kinds of nodes sensor left and general right connected by an edge from the sensor to the general one At any time every node has twos activity parameters called real and imaginary states These states may vary from zero to one where zero means no activity and one means full activity The way the real state is calculated depend on the kind of node being used each of them has a specific activation function that will always depend on the activity coming from other nodes or inputs The only kind of nodes that changes their activity based on the inputs is the sensor node they simply copy the environment that is generally speaking the input given by the user All the implemented types of node in relation to their activation function will be described in the Activities section The expression for the real activity of a given node k over time is described by 19 r t A I t where A is the activation function of node k and I is the vector of size n containing the n inputs to the node k Imaginary activity of node k over time i t P PI t where P is the prediction function and PI t is the prediction input vector calculated from the states of the nodes at the outputs of k at time t Node id
43. fected by this fact and gives rise to the concept of level The levels work as a hierarchy meaning that every node correspond to a higher level than all the nodes it receives information from This can be seen as levels of abstraction since the higher the level is the more concept have to be active and associated The levels are labeled with increasing numbers starting at 0 which correspond to the sensors level and up to the highest level where nodes have no outputs Level 0 Level 1 Level 2 Sensors 0 A AAA A AS 0 42 0 24 042 0 85 0 85 p 1 es A i 0 85 0 85 085 0 85 Figure 3 Example of how the nodes are shown by levels in the toolbox and how to interpret them Inhibitory edges This kind of edge inhibits the activity of a node depending on the state of a set of other nodes The inhibition activity over a node will corresponds to the addition of activities in the nodes inhibiting it that addition is truncated at a maximum of 1 ensuring the inhibition over a node fits in the interval 0 1 The inhibition is performed after the activation has been calculated by multiplying the complement of the inhibitory addition When inhibition is applied the real activity of node k over time is modified by ret A I t 1 h t where h t is the sum over all the inhibitory inputs of k at time t In the interface this kind of edges is depicted as orange connection between nodes 21 0 85
44. fferent fields and which solutions imply applications of huge relevance In order to contextualize the concepts associated to this model an introductory description is presented depicting some approaches that aim to cope with problems similar to the ones faced here These approaches correspond to some ideas and models emerged throughout the development of the fields of cognitive modeling problem solving and robotics and that are of relevance for understanding the challenges and needs addressed by this research project Background The challenge of creating systems capable of mimic reasoning and cognition has been addressed by many and from different perspectives and disciplines There are some proposals broader than others but the main goal has been mainly related to the ability of creating concepts and manipulating them in order to draw conclusions and deliver responses Most of the approaches could be classified in relation to the way information is organized and processed the general division usually is into emergentist symbolic and statistical or probabilistic models Among these there are certain conceptual differences that give advantages and disadvantages to each as will be shown In the mentioned classification the division into Symbolic and Emergentist also called non Symbolic approaches is broadly used in the field They basically difter in the way they create represent and manipulate concepts and knowledge The symbolic perspe
45. file Tool bar in the tool box New network Ctrl N Creates a new network of general purpose that depending on settings and the input stream with which is fed can be used in manual or interactive mode Open network Ctrl O Loads a previously saved network in a TNN file Save network Ctrl S Saves the network in a TNN file in a specified path if no such a file has been specified it will open a file dialog in order to select it Open input file Ctrl 1 Opens an input file with which the network will be fed it looks for text files txt Add nodes Ctrl A Opens the add node dialog with which the nodes to construct the network can be added Some of these functions are found in the File menu plus the function Save network as which allows to change the destination file at which the network is saved lt also includes the Exit item Ctrl Q Edit Help New Network Ctrl N Open Network Ctrl 0 Save Network Ctrl S Save Network As Open Input File Ctrl I Exit Ctrl Q Figure 11 The file menu in the tool box 39 When add nodes is called the following dialog is displayed allowing to choose the kind of node to be added It displays a description of the function the selected kind of node performs and the quantity of nodes to add can be selected 12 Add Nodes Node types Input Group Max Node Average Node S Buffer Node Delay Node Complement Node Sigmoid Node Simple Gaussian Node Compl
46. hen creating new types Attributes Nodes It is a list of objects of the class Node where the index of each element corresponds to the id of the corresponding node Levels It is a list containing lists of nodes Each list of nodes is a level and points to the nodes corresponding to that level The level id corresponds to the index in the main list inputNodes It is a list containing the ids of all the input nodes inputGroups Is a list containing lists of ids each list contains the ids of all the input nodes that belong to a group The id of each group corresponds to the index in the main list associationNodes Is a list of nodes of the class AssociationNode that points to all the nodes of this class is used to keep track of the existing associations and check the existence of a particular one when looking for new possible ones emotionallmpact Is a field updated at every time step and is used for control it is the one used as recording signal in the interactive mode and always takes the value of the first position in the input array lastEmotionallmpact It retains the value of the emotionallmpact at the previous time step and is updated when a value is set for emotionallmpact depth It s an integer that represents the number of levels the network has associationThreshold It is used in the interactive mode lt is a number in the interval 0 1 that indicated the minimum reverberation a node
47. his sense the TNN model lacks crucial characteristics as it cannot interact with its environment at all and actually its development is not related to any kind of interaction beyond the inputs it receives In fact interaction may be needed to achieve the automatic characteristics that would remove the need for the control signals that are implemented in the current model Nevertheless the TNN model is still being developed and further versions of it may take into account lessons from the present work and existing models and approaches that have faced similar problems Therefore as a consequence of the discussed issues and ideas some particular future work considered relevant for the TNN is mentioned in the following section 55 Conclusions The toolbox implemented successfully satisfied the needs and met the requirements under the constraints given by the partial development of the TNN model Basic inferences inductions and generalizations achieved are linked to the symbolic manipulation of the input grounded on previous knowledge of the problem by the user reason why there is not enough information or evidence to claim that in general the model is capable of the two main reasoning capabilities aimed The emphasis on symbolic meanings of the inputs in the problems definitions may be restricting the model into becoming completely symbolic one The model still requires of great deal of control signals and thresholds definitions
48. ighlighted before the concept is still being developed and so are the algorithms and implementation details then all the results reported here are the outcomes of an iterative process that led both the concepts and the implementation So and as the research is to keep on advancing the following description focuses in the usage and the concepts included as well as in technical details that are considered necessary for further development of the tool The building blocks of a TNN The networks are built using basic elements that are related to each other and possess specific information that allows the network to work These elements or building blocks in TNN are denominated Nodes and Edges Nodes The TNNs as implied in the name are the interconnection of a given number of elements that share information Thus it could be seen as a directed graph but in this case it deals with two kinds of information that flow in opposite directions as will be explained later These interconnected elements in the network are called nodes they are individual processing units that can be selected and added to the network either manually by the user or automatically by the tool when it is specified to do so All the nodes in the network represent simple concepts that can be labeled by the user in order to keep the transparency In a sense of conceptual learning every node represents a concept that basically comes from the association of either previously
49. ist are largely focused on inductive learning d Avila Garcez and Lamb 2011 Therefore one of the major goals of TNN is to include both deductive and inductive reasoning as simultaneous capabilities of the same model Of course there have been many different attempts to achieve that with hybrid architectures nevertheless the fundamentals of TNN differ in the stress on the need for models that remain transparent while achieving the two kinds of reasoning by means of just one process It has been common that when merging approaches the architectures tend to have different structures for symbolic and sub symbolic processing that are connected but still independent In the case of TNN the proposal is to achieve both the deductive and inductive capabilities by means of a single structure and a single learning algorithm As mentioned the main goal with the structure proposed in TNN is to keep the transparency also called interpretability which refers to a model being easily understood or interpreted by its users This fact is stress as the problem of lacking transparency is an issue that affects many models and especially those 15 based on connectionist approaches which leads to great problems when interpreting and grasping the underlying process of a structure even if it solves a particular problem An example of that are the feed forward artificial neural networks in these structures there is not much transparency since it is not trivi
50. ity of the elements was achieved by means of a simple object oriented approach that after deployed allowed a very fast editing which represented an advantage for experimenting with many changing ideas and designs during the development of different concepts of the TNN The final interface fulfilled completely the desired characteristics for the application and even went beyond the requirements The usability of the tool is based on a very simple and intuitive interface that once the concept of TNN is clear allows creating training and manipulating a network in about three simple steps The final result offers a great deal of freedom to get information of the elements of the network as well as to manipulate and customize them stressing the point of transparency on which the whole project is based The ability to move freely through the network by just one click and as the zoom is easily manipulated exploring results into a very simple and helpful task especially when the networks grow to some many nodes In general the feedback that the platform allowed while it was being improved permitted highlighting different drawbacks of presented approaches for the TNNs and opportunely fostered new proposal to improve the model 52 Discussion As implemented the TNN model allowed performing basic inferences inductive learning and generalization for specific problems as depicted in the results section All of the capabilities are achieved keeping
51. k and a bad selection can easily lead the network to an explosion of redundant associations Redundant and unnecessary associations certainly make the network not really useful and affects the transparency as it reaches states at which the meaning of the concepts created are incomprehensible To give an example of this if a sequence of four sensors is shown repeatedly to a network and the association threshold is set too low in this case 0 5 even when there is control by means of the recording signal there are at least three associations created for this sequence in the first level Each of these associations is allowed by the constructions rules as they have different inputs All of these associations get activated as the sequence is presented again and therefore will be associated again at the next level This process is repeated over and over again as the sequence is repeated creating an endless number of levels as depicted in figure 27 0 5 0 09 Figure 27 Explosive growth in lack of proper control This drawback on the controllability affects the utility of the model and its own transparency creating a need for new approaches in relation to the construction rules 51 On the Toolbox In relation to the main objectives set for the thesis in terms of the implementation the toolbox turned out to be a very beneficial and easy to use instrument for the developing of the TNN The goals of modularity and flexibil
52. late many nodes 26 As mentioned the simultaneous activation of a node is just a particular case of the temporal relationships at which all the activations are presented in the same time step In this kind of associations the order does matter for example given two nodes a and b that belong to the same level may have two possible temporal associations this is If node b gets activated after node a got activated it is a different association than if a gets activated after b On the other hand differences in time are not considered as different associations lf the order is the same that means that if b gets activated one time step after a it will be considered as the same association than b getting activated two or more time steps after a did so The activation of an association must represent how much of the actual relationship is achieved which implies that in the case of temporal relationships the activity must relate different time steps To achieve that the activity reverberation of the edges going to the association and in particular their decay parameters are used to enclose temporal information The activity reverberation of an edge copies the real states of the node it comes from and decays depending on the decay parameter this parameter is to be learnt by experience and has to do with how many time steps the whole association takes to be complete after the node sending information through the edge was first activated The
53. lessly spontaneously and with remarkable efficiency The project attempts to show how a connectionist model can be capable of encoding semantics systematic mapping and knowledge about entities and also be available to perform reflexive inferences in a fast and efficient manner This is done by creating structures that represent schemas by focal cells clusters and generating inferences by the propagation of rhythmic activity over those clusters Thus all information processing is based on temporal synchrony throughout a structured neural representation This fact is claimed to demonstrate how such a connectionist structure is sufficient to achieve rational processing in the brain This model is related to different projects related to decision making problem solving and planning and language acquisition The International Computer Science Institute 2012 In general connectionist models are capable of simplifying and generalizing data from complex inputs to more reduced spaces in the way of inductive learning Also some connectionist models have been merged with other approaches to achieve better capabilities as will be described further below 8 Probabilistic approaches Probabilistic models can be classified mainly as top down approaches that relate concepts and perform selections depending on probabilities learnt through experience The most basic and classic yet relevant example of this are the Bayesian Networks Probabilistic mod
54. lop intelligent systems then the model has to be focused more in the creation of concepts and behavior based on interaction and not only on extracting information from the inputs Thus a more context based and interactive learning both for the model and the implementation is suggested for a better and more interesting progress 58 References C Stranneg rd O H ggstr m J Wessberg C Balkenius 2012 Transparent Neural Networks paper presented at the SweCog C Stranneg rd 2011 Transparent Neural Networks manuscripts March 2011 Chalmers University of Technology Troy D Kelley 2003 Symbolic and Sub symbolic Representations in Computational Models of Human Cognition What Can be Learned from Biology Theory 8 Psychology Vol 13 No 6 2003 pp 847 860 James L McClelland Matthew M Botvinick David C Noelle David C Plaut Timothy T Rogers Mark S Seidenberg and Linda B Smith 2010 Letting structure emerge connectionist and dynamical systems approaches to cognition Trends in Cognitive Sciences Vol 14 Issue 8 August 2010 pp 348 356 Thomas L Griffiths Nick Chater Charles Kemp Amy Perfors and Joshua B Tenenbaum 2010 Probabilistic models of cognition exploring representations and inductive biases Trends in cognitive Sciences Volume 14 Issue 8 August 2010 pp 357 364 Nick Chater Joshua B Tenenbaum and Alan Yuille 2006 Probabilistic models of cognition Conceptual foundations Trends in Cog
55. lue of all the imaginary activities of the nodes at its outputs multiplied by the respective weights The imaginary activity is calculated by izlt P PI t max W1 t ig t g 1 n where W1 t is the weight from the node g which is at an output of node k and i t is the imaginary activity of node g In the example on Figure 5 the node at the deepest level copies its real activity as its imaginary one but meanwhile the nodes on the previous level have slightly different weights and therefore different imaginary activity that depends on the node at the deepest level The sensors have weight close to 1 0 and as described they take for imaginary activity the maximum of the possible activities coming from nodes at their outputs 24 0 2 s4 Pre gt q 3 1 1 ee Bm e 0 47 0 0 0 46 0 5 Figure 5 A simple network at a given state to depict the real and imaginary propagation The node at the deepest level copies its real activity as imaginary while in the others the imaginary activities are affected by the whights Time step Updating the states of the network means to take an input and calculate the real and imaginary activities of all the nodes in the network The update of the network is carried out every time a new input is presented this is defined as a time step In this sense there is no delay in between the moment the input appears and the moment the activity of all the nodes are updated That means as w
56. mainly concerns with merging characteristic of very heterogeneous systems such as the symbolic and connectionist models These two approaches have very different types of representation learning and processing therefore most of the proposals are architectures that attempt to use symbolic perspective for manipulation and connectionist approaches for learning In other words a top down system that is fed by a bottom up one Troy D 2003 A way of seeing this proposed by Troy 2003 is that cognition can be considered as a cognitive continuum with two ends at a highest end the symbolic processing is carried out which could be interpreted as the equivalence of the prefrontal cortex in the human brain At the other end of that continuum the lowest level is related to the most basic input processing which in the human system could be equivalent to the reflex nerves But still the link between the two ends of that continuum is not yet clear Hence though in hybrid architectures the sub symbolic systems present favorable issues related to learning the symbolic processing is still mainly related to representation and inference which transfers many of the symbolic systems flaws to the hybrid structures Sun 2001 This means that at the symbolic level the structures are still highly dependent of a knowledgeable user and therefore not much is really left to learning through experience Troy D 2003 Nevertheless the usage of various approache
57. mean learnt and o is the standard deviation Complete Gaussian Node s a compilation n of function like the one described for the simple Gaussian node where n is the number of inputs of the node and individual parameters are learnt for each of them The final result is the multiplication of all of these functions n 1 41 A t e 2 i 1 Sensor Nodes that set their activity from the input given by the user environment 23 The second kind of activity is the imaginary this is meant to infer or complete information from the one present at a specific moment in the input however imaginary activity can also be a prediction of information over time as a relation to expected concepts or inputs in both the past and the future The imaginary state of a node will depend on the state of those that are fed by its real activity Nonetheless when a node does not feed any other or has no outputs then it will copy its real activity as imaginary activity in order to use it as the source for inference To calculate the imaginary activity of a node weights at its outgoing connections are to be learnt After the proper learning the value of these weights corresponds to the probability of the node being active when the node at that output is active Every node that has connections going outwards adapts a weight for each of those edges Then when the imaginary activity is being propagated the imaginary activity is set to the maximum va
58. n The activation functions depend on the goal of the node and the way the information is fed to it the information can be the real activity of the nodes that are feeding it or the activity reverberation in its incoming edges reverberation activity The types of nodes implemented regarding their real activation function is listed here 22 Min Nodes Nodes which activity is set to the minimum at its inputs A I t minfI t i 1 n Max Nodes Nodes which activity is set to the maximum at its inputs A 1 t max i t i 1 n Average Nodes Nodes which activity is set to the average of its inputs A I t mest dy Delay Nodes Have only one input and set their state as the real activity at its input in the previous time step r t It 1 Size of vector l is always 1 Buffer Nodes Have only one input and copies the same state that the real activity at its input used to bring the same activity to a deeper level r t I t Size of vector l is always 1 Association node Average of the real reverberation activity at their incoming edges n Blt A I t Ai Ett where B t is the reverberation vector in the inputs of node k at time t Simple Gaussian node Have only one input and learns by experience the average and the standard deviation of the inputs shown The real activity is calculated by means of the parameters learnt using a bell shaped function I y A I t e 20 where u is the
59. n hybrid approaches at that level An example could be to focus on more behavioral models aside of the cognitive perspectives or as their complement in a similar way as they have been opponents and complementary approaches in psychology Similarly a very relevant approach to the goal on general problem solving that leaves aside the constraints of architectures aiming for specific tasks is the developmental or epigenetic robotics described below Epigenetic robotics The goal of epigenetic robotics also known as developmental robotics is to model the development of cognition through the usage of elements from different sciences and approaches such as robotics neurophysiology psychology and artificial intelligence where the results may be a beneficial exchange among all of them Metta Giorgio and Berthouze Luc 2005 This is carried out by the study of the development as a process in which modifications on cognitive structures lead to an overall emergence of abilities which in human basically happens form the embryo to the fully developed adult Here development is seen as an open ended adaptation process generated by means of interaction with the environment Metta Giorgio and Berthouze Luc 2005 The whole idea emerges from the need across the cognitive sciences for models that can scale up beyond specific domains and scenarios and that at the same time can display a developmental trajectory and are transparent in their constru
60. nductive reasoning at the same time The transparency on one hand is a point that has been stressed during the implementation and was explained in the corresponding section of this document On the other hand the performance of the model on the proposed abilities for reasoning has not been shown directly so far Therefore this results section is mainly focused on showing how this model deals with these kinds of reasoning through some basic examples Descriptive examples Simple associations This example shows how a simple association is created when three inputs appear simultaneously and how partial information elicits certain prediction in form of imaginary activity At the first time step of the example the three inputs are completely active as well as the recording signal At the following step the inputs all go down to 0 and the association is created Node 3 in figure 20 Complete pattern Partial information Figure 20 Depicts the differences betwwen the activation of an association node when the information is complete and when it is partial This simple example can be used to show how the imaginary activity of the nodes is useful to predict or infer possible inputs out of partial information To show this the network is fed with partial information taken into account that in this example there is only one association the predictions of the sensors will only depend on that association 44 If the input activates
61. nitive Sciences Volume 10 Issue 7 July 2006 pp 287 291 Amanda J C Sharkey 2009 Artificial Neural Networks and Cognitive A Modelling Encyclopedia of Artificial Intelligence 2009 pp 161 166 Inc Numenta 2011 Hierarchical Temporal Memory including HTM Cortical Learning Algorithms The International Computer Science Institute 2012 shrut online Available at lt http www icsi berkeley edu shastri shruti gt Accessed April 2012 Bringsjord S 2008 Declarative Logic Based Computational Cognitive Modeling in Sun R ed The Cambridge Handbook of Computational Psychology Cambridge UK Cambridge University Press 2008 pp 127 169 Lewis R L 1999 Cognitive modeling symbolic In Wilson R and Keil F eds The MIT Encyclopedia of the Cognitive Sciences Cambridge MA MIT Press 1999 R Sun 2001 Artificial intelligence Connectionist and symbolic approaches In N J Smelser and P B Baltes eds International Encyclopedia of the Social and Behavioral Sciences pp 783 789 Pergamon Elsevier Oxford University of Michigan 2012 SOAR online Available at lt http sitemaker umich edu soar home gt Accessed April 2012 59 Perlovsky L I 2007 Neural Dynamic Logic of Consciousness the Knowledge Instinct In Eds L I Perlovsky R Kozma Neurodynamics of High Cognitive Functions Springer ACT R Research Group Department of Psychology Carnegie Mellon University 2012 ACT R online A
62. o any other value incomingEdges It is a list of objects of the class Edge that represent the edges going to the node outgoingEdges Is a list of objects of the class Edge that represent the edges going out of the node depth It represents the level in the network at which the node is lastinput It is an array containing the values received as input in the last time 35 step lastPredictionInput It is an array containing the values received as input for updating the imaginary state in the last time step Abstract methods double updateStateFunction double statesOfinputNodes This function takes as parameter an array which must contain the information from the nodes at the inputs of the given one that are to be used to calculate the state The order of the states in the array is assumed to have the same order that the nodes have in the incomingEdges list The function must perform the calculation corresponding to the node type and return the result in the interval 0 1 void trainFunction double relevance This is used if the node needs to update any parameter after the update is performed at the end of every time step It receives as parameter a number in the interval 0 1 that may be used to indicate the relevance of the example being trained in case of being necessary In the General Network the only kinds of nodes that train values through this function are the simple Gaussian and the complete Gaussian no
63. ome psychology experiments Sharkey 2009 However this kind of experiments are limited to a specific task and even when the data is fit it cannot be said that the networks mimics the process itself or even more risky that it is comparable to actual reasoning in fact it is really hard to interpret the actual behavior in the network that leads to the result but it is known that at the end it performs nothing but an error minimization task Moreover traditional neural networks are dynamic systems that can accomplish very good performance on optimization and data fitting This is why most of the development on this field has been done aiming to solve particular problems hard for traditional mathematical optimization methods actually most of the variations of the artificial neural networks have emerged to fit particular optimization problems But when it comes to the ability of modeling cognition or actual reasoning based on them it is not so clear that these structures possess it Nevertheless artificial neural networks are not the only connectionist model some other models have been proposed based on connectionist ideas specially aiming to create the ability of learning concepts and use them for inference An example of a connectionist model is the Shrut architecture The International Computer Science Institute 2012 Shrut is an architecture that focuses mainly on drawing inferences which its authors proclaim to be performed effort
64. onal groups of symbolic entities that may in turn contain or be contained by other groups A well known of these representations are the semantic networks these networks mainly used for language representation and processing are formed by labeled nodes representing concepts related in turn by labeled links Sun 2001 An example of a symbolic architecture is the SOAR project University of Michigan 2012 lt is a production rule system where problem solving is basically related to representing a search space All the decisions are taken by using the interpretation of the sensory data and the compilation of relevant knowledge coming from previous experiences The main goal is basically to create a general problem solver where every problem regardless of its characteristics is defined as a search space University of Michigan 2012 On the other hand many symbolic models have used logic as a representational language For many the idea of logic has always been a part of the supreme goal of developing intelligent machines the promises back in the 50s was on using logic as the mechanism to build computational artifacts available to even exceed human minds in terms of intelligence The truth after many years of research is that the fundamental ideas based on formal logic have failed to accomplish the task Nevertheless people still believe in logic and it plays a fundamental role on many algorithms and as mentioned before is used to analyze perf
65. only the second sensor then the association will be active at approximately a 33 and as the only examples shown to the network has been the complete association including the three sensors the prediction to each of them is going to be a third as well Note that if the pattern is complete the prediction at every sensor will be practically 1 Composed concepts and inferences The following example shows how a network that has created a composed concept out of two previous ones In this case two different concepts are shown to the network separately each of them relates two specific sensors Afterwards the two concepts are shown at the same time and that creates another at a deeper level When each concept is presented at different time the network creates an association node for each of them called concept 1 and concept 2 in figure 21 Afterwards when the two concepts are shown together the network uses the associations created previously to build a composed concept at a deeper level representing the two basic ones together That means that no concept is created including the four sensors since deeper composed concepts are preferred by the construction rules This can be seen as an inductive learning since the network is creating more abstract and concrete associations as it finds relationships in the activation of more basic ones Concept 1 Composed concept Concept 2 Figure 21 Depicts how to interpret the formation of a con
66. orks model to be introduced This is the more general question on how intuitive the fundamentals of cognition or reasoning could really be As mentioned by Chater Tenenbaum and Yuille 2006 people struggle not only with probability but with all the branches of mathematics and this does not changes the fact that for example as hard to understand as it could be Fourier analysis is fundamental in audition and vision in biological systems Therefore it may be sound to state that analyzing the complexity behind the model or its easiness of interpretation may not be the best choice to compare performance But regardless of which could be the best measure it is also undoubtedly relevant to identify biases in the measures that may be favoring particular interpretations of reasoning or cognition when proposing or evaluating a model As the field has been broadly focused on architectures and structures such as those in connectionists or rules based models the performance measurement has to do with their characteristics and the ideas behind them such as logic and heuristics Griffiths Chater Kemp Perfors and Tenenbaum 2010 This kind of biasing in the analysis may of course affect models that are based on different perspectives such as the probabilistic models However regardless of the 9 models that it could benefit or affect this kind of biasing may lead to extremely dangerous assumptions an example of this is treating logic almost as an
67. ormance in different kinds of models Perlovsky 2007 Many go back to Aristotle to describe logic and argue how even from syllogisms intelligent responses can be described The main element of this argument arises from the idea that Aristotle inferred that certain context independent structures can describe and predict human thinking However it has to be also taken into account that these studies on logic came from the search of a supreme way of argumentation but never as a theory of mind Bringsjord 2008 Thus to talk about logic as a model that describes human thinking may be too broad in the sense that argumentation is a particular characteristic of language and its consequence But the fact that language and after it logic emerge from reasoning and thinking does not necessarily mean that they are the basis from which those phenomena emerge In other words arguing that a consequence can be also the basis of the same process may lead to some partial understanding of the problem Nevertheless approaches based on logic have been successfully implemented to solve particular problems often related to symbolic manipulation Many problems in language processing have been addressed by these approaches complemented in many cases by probabilistic models Other broadly explored area is the one concerned with constrain satisfaction problems which are addressed by logic based models and particularly bi valuated systems Logic approaches have al
68. pulation of basic structures and elements It considers characteristics of the developmental robotics and cognitive modeling Taking all of these concepts and goals as a basis the main objective of this thesis is to help in the development of that model by implementing a toolbox that allows the creation and evaluation of the networks leading to conclusions and meaningful feedback that fosters proposals for further development on the model Contents TIPO GUC OR POCO AO ale eee acre cen auth rae E tari 6 Background ii sd 6 Transparent Neural Networks 00 ia 15 The implementation oo so as eae as a Na ace ea a reenen 18 The building blocks ofa TAN cas aldo 18 Networks construction modes ciisediinscntagnases ontario 26 Network Working CV CIC cuts arta cuida 30 General description of the implementation ooconnincccccinnnnnnnccccccncnnnanancnninos 33 PRES CUTS can 44 Descriptive examples ci a 44 ORT TOD ears ce cece nor an nep 52 DISCUSSION a debes 53 CoOnclUsiOA Sie 56 FUTUES WOTK erecto isis 57 o eit alsin fae Rist Caleta triste lel al Ai ih heel Rial 59 Introduction The goals of the Transparent Neural Networks project to which this thesis is directly connected are related to achieving higher cognitive functions such as deductive and inductive reasoning as well as automatic learning by means of transparent and interpretable structures But all of these objectives are in turn related to problems that have been studied in di
69. rnt through experience From the very beginning of formal computation and the first ideas on Al symbols and specially logic was considered as a basic mechanism by which minds work The idea is that symbolic representations stands at the very core of how intelligence work and therefore the focus is set on what symbolic knowledge an agent would need in order to behave intelligently Bringsjord 2008 Then this perspective focuses not on how the knowledge arises but on how it should be used The symbolic approaches envision cognition as some sort of computer programs and describe aspects of cognition and their emerging results as a set of basic computational processes claiming that this idea could produce for example predictions with performance comparable to humans Lewis 1999 A foundation of this approach is the so called physical symbol system hypothesis proposed by Newell and Simon lts idea is to use basic symbols as representational entities combine them to form expressions and manipulate those expressions to create even new ones Their claim stated that A physical symbol system has the necessary and sufficient means for general intelligent action and is an idea that has been the foundation of massive efforts in research in Al Sun 2001 10 Many symbolic representations aim to capture and organize knowledge in the form of structures or architectures The idea for the structures is to organize knowledge by creating relati
70. robabilistic in fundamental ways Chater Tenenbaum and Yuille 54 2006 Thus it is questionable to try to emerge intelligence from a fact that is not completely related to the actual phenomena being modeled Nevertheless this partial definitions on intelligence are common to many models for instances in the SOAR architecture University of Michigan 2012 the ultimate goal in intelligence and complete rationality is settled as the ability to use all available knowledge to solve any problem the system encounters but then again if rationality is inspired by human behavior the question is why such a crucial definition does not take into account the fact that humans never consider all the possibilities when taking a decision but just some particular ones that depend on parameters of which we may not be even conscious at all Overskeid 2008 On the other hand there are approaches such as the epigenetic robotics that emerges from the need of robots to understand and develop in relation to their environments and rejects more classical views of robotics in which the capabilities of robots are completely based on pre programed behaviors that removes any possibility of concept creation and development This approach also states the absolute need for the robots to have a body with which to explore and verify knowledge which implies that any model to develop knowledge and intelligence must be available to interact with the environment In t
71. s certainly enhances the capability of the models Several architectures have been developed based on hybrid structures and some of them have achieved reasonable results and are known as relevant cognitive architectures among them ACT R is a typical example ACT R ACT R Research Group Department of Psychology Carnegie Mellon University 2012 is an architecture that is born with the goal of understanding human cognition and how knowledge is organized and used to produce intelligent behavior This architecture has been evolving for many years reaching interesting results in various fields related to cognition This architecture has been used by researchers to produce data on theories that can be directly compared to experiments with human participants This allows verifying models on cognition directly by means of the architecture Some of the models created with ACT R include learning and memory problem solving and decision making language and communication perception and attention cognitive development and individual differences ACT R as a hybrid architecture has both symbolic and a sub symbolic structures the symbolic one is a production system that matches the state of the system to previously learnt symbols The sub symbolic structure is a set of parallel processes that control many of the symbolic elements through a series of equations and in many cases in accordance with utility functions ACT R Research Group Department of Ps
72. so evolved during the last decades to allow more flexibility than the formal logic which as limited to discrete truth values runs into troubles easily Just in 1902 Russell showed a whole in formal logic which caricature is described by this simple example A barber shaves everybody who does not shave himself Does the barber shave himself The Cambridge Handbook of Computational Psychology 2008 pp 127 169 Any possible answer to this problem yes or no is contradictory Problems like that and later more complex ones led to the rise of concepts such as multivalued and fuzzy logic where variables can take many values or virtually any value in an interval between the classical true and false These more flexible approaches have allowed addressing a broader range of problems but have also shown the need to merge logic and other approaches to achieve better results as described in the next subsection Hybrid models It seems reasonable to aim for a model that includes both top down and bottom up ideas as they can be complementary For example symbolic approaches that are mainly concerned with deductive reasoning may be complemented by connectionist approaches that are mainly focused on inductive learning d Avila Garcez and Lamb 2011 Thus it is easy to advocate for the search of such a model and indeed that is not a new idea nonetheless as easy to argue about its reasons not so easy is the task of developing it A hybrid model
73. symbols to solve particular tasks When analyzing the existing models for cognition and problem solving one may have the sense that generally all the applications aim to solve a particular task that the researches have in mind and leave many details apart This fact is reasonable as the goal is based on solving specific problems however it is a very narrow perspective if the goal is to enhance the performance and capability of models or aiming for a more general problem solving approach The fact is that as mentioned by Ekbia 2010 there is a utilitarian notion of human life as being composed by a set of problems and human intelligence as nothing but a capability to solve them The issue with this idea is that it somehow neglects that the human brain and in fact any other brain even when capable of solving problems by sequences of steps is rather a dynamic system with many structures shaping behavior and the basis of its characteristics should never be confused with that particular ability of describing problems by sequences or by any other semantics For example when one focuses on an specific problem and asses intelligence based on the ability of performing clear steps for reaching a desired solution one must also think that in reality humans do not always reason in a correct way Bringsjord 2008 In fact psychological works by Kahneman Tversky and colleagues suggest that human cognition might be non rational non optimal and non p
74. the networks are carried out On the other end of the diagram is found the Graphical user interface which deals with all the graphics generation and interaction with the user as well as the information flow between the user and the toolbox And finally to manage the link between these two main blocks there is an interface that deals with the communication and information flow between the network and the interface Graphical User Interface Network As the focus of this work is on the design of the network the description below is focused only on the main block Network therefore the descriptions regarding the implementation for the Graphics and the Interface blocks are not included The whole system is created under an object oriented paradigm and for the Network block basically there are three principal classes the Networks the Nodes and the Edges The Node and the Network classes are abstract classes and the different type of nodes and networks are classes that extend the main ones implementing the abstract methods that differentiate them These classes are created abstract in order to allow future implementation of new kinds of nodes or networks However in the current description the only network used is the so called General Network which is designed to work in both interactive and manual modes The figure 9 depicts a general class diagram where the main relationships and inheritances are shown There appear all the node kinds that
75. ting is performed after reading the corresponding input array and copying it to the sensors as real activity then the process of updating starts First the update of the real state is performed by propagating the real activity forward The second step is to update the imaginary activity going backwards Once both the activities have been propagated the recording signal is checked in order to decide whether to look for associations or not and in any case the last step is always to perform training Update Real States Update Imaginary activity Look for Recording Associations signal Input format The input for each time step must be an array of size n 1 where n is the number of sensors the network has The first element of the array must be the recording signal which is to be different to zero only if the network is expected to create associations The rest of the inputs correspond to the value of the input sent to the sensors In the network each sensor has an ld which corresponds to the order in which they were added these ids are the order used to update them from the input array If the size of the input is shorter than n 1 the inputs given will be used to update sensors from the first id until the end of the array 30 Reccording signal Input array N ON a 10 5 0 94 0 25 0 75 0 12 Figure 8 Input format Update Real states When updating the real states all the levels are checked starting at level 1 since lev
76. vailable at lt http act r psy cmu edu gt Accessed April 2012 Ramamurthy Uma Baars Bernard J D Mello Sidney K Franklin Stan 2006 LIDA A Working Model of Cognition The 7th International Conference on Cognitive Modeling Trieste Italy April 2006 Eds Danilo Fum Fabio Del Missier and Andrea Stocco p 244 249 published by Edizioni Goliardiche Trieste Metta Giorgio and Berthouze Luc 2005 Epigenetic robotics Modelling cognitive development in robotic systems Cognitive Systems Research Volume 6 Issue 3 pp 189 192 Anthony F Morse Joachim de Greeff Tony Belpeame and Angelo Cangelosi 2010 Epigenetic Robotics Architecture ERA IEEE Transactions on Autonomous Mental Development Vol 2 Issue 4 December 2010 Stoytchev Alexander 2009 Some Basic Principles of Developmental Robotics IEEE Transactions on Autonomous Mental Development Vol 1 Issue 2 August 2009 Asada Minoru Hosoda Koh Kuniyoshi Yasuo Ishiguro Hiroshi Inui Toshio Yoshikawa Yuichiro Ogino Masaki and Yoshida Chisato 2009 Cognitive Developmental Robotics A Survey IEEE Transactions on Autonomous Mental Development Vol 1 Issue 1 May 2009 d Avila Garcez Artur S and Lamb Luis C 2011 Chapter 18 Cognitive Algorithms and Systems Reasoning and Knowledge Representation Perception Action Cycle Models Architectures and Hardware Models Algorithms and Systems Springer Series in Cognitive and Neural Systems
77. ychology Carnegie Mellon University 2012 ACT R shows many of the advantages of the hybrid models and in fields as applied psychology it has grown interest on more integrated cognitive architectures However it still exhibits deficiencies typical to these architectures for instance and maybe the most important one most of the knowledge acquired depend completely on the programmer and not on learning from the environment Troy D 2003 Another known cognitive architecture that uses a hybrid approaches is the LIDA architecture LIDA uses both symbolic and connectionist approaches merged together The architecture is based on a cognitive cycle that goes from perception to action During this cognitive cycle several aspects are taken into account but always with special emphasis on the roles of feeling and emotions Emotions are used for conceptualization and are related by associative relations they guide actions and what is called consciousness in the model which affects decision making at every level Ramamurthy Baars D Mello Franklin 2006 As these two examples many others have also shown that in general merging approaches is a feasible way to improve performance and that keep on generating new points of view on the overall problem solving goal However most of the approaches have been based on a cognitivist perspective but it 13 does not mean that it should be the only one it could be reasonable to evaluate others or eve
Download Pdf Manuals
Related Search
Related Contents
DCS400 MANUAL 071009 4P:ps - Northern Tool + Equipment Olympus C-570 Digital Camera User Manual 5890dn - Support Sagemcom User Manual UH Murphy Series LM300 User's Manual 取扱説明書 - M DIO-3232 Digital I/O Card User`s Manual (V1.5) sfigmomanometro ambulatoriale clinic digital sphygmo Copyright © All rights reserved.
Failed to retrieve file