Neuro-Cognitive Architectures

Knowledge Consolidation and Inference in the Integrated NeuroCognitive Architecture Richard J. Oentaryo, Nanyang Technological University, Singapore Michel Pasquier, American University of Sharjah, United Arab Emirates

The Integrated Neuro-Cognitive

D

onald Michie once argued that AI is about “making machines more fathomable and more under the control of human beings, not less.”1

This is worth remembering, given that current technological progress is

Architecture (INCA)

rendering our environment ever-more incomprehensible. Such a paradoxical

emulates the putative

situation results from engineering more sophisticated hardware and software

functional aspects

tools and systems that fewer can master, apply, or improve. Simplicity is much needed, but achieving it is a grand challenge. Hence, there is a need for more general machine intelligence that can provide the necessary natural interaction and human-like cognition. Recent advances in cognitive neuro­ science, psychology, and AI provide remarkable insights toward building an integrated framework for human-level, general intelligence. It will likely be achieved by leveraging AI technologies in innovative ways while incorporating novel ideas, from biologically inspired models to human factors and cognitive engineering. The ultimate aim is to create autonomous systems that can learn and operate in varied domains and solve complex problems in a human-like manner. Such systems would help make machines

of various major brain systems via a learning memory modeling approach.

62

easier to build and use, which is a crucial goal of AI. By extension, they would also let us address many critical issues in complexity, security, robustness, maintenance, and usability of computer-based systems, on which the world’s key infrastructures rest today. The ongoing quest toward an integrated theory of machine intelligence has led to the investigation of many cognitive archi­ tectures, 2 –5 or generic blueprints for building intelligent agents, integrating and testing computational models of knowledge representation, reasoning, learning, and so on. Their main characteristic is to offer a holistic framework for general intelligence that works across different task domains, in contrast to narrow AI techniques that focus on specialized solutions to welldefined problems (such as expert systems).

1541-1672/11/$26.00 © 2011 IEEE Published by the IEEE Computer Society

IEEE INTELLIGENT SYSTEMS

Developing Cognitive Architectures

S

tudies in cognitive neuroscience and psychology have widely established that the most fundamental features in developing any cognitive architecture are memory and learning.1 Memory denotes the repository of background knowledge about the world and oneself, while learning is the key process responsible for shaping this knowledge. Together, they form the indispensable substrates for higher-order intelligence, such as decision making, planning, executive regulation, and creativity. A simple taxonomy based on the memory and learning properties categorizes existing cognitive architectures into three groups: symbolic, emergent, and hybrid. 2 Symbolic architectures, such as the state, operator, and result (SOAR) rule-based architecture,3 are generally built upon a top-down, analytical AI approach operating on abstract physical symbols or declarative knowledge. Typically, these architectures involve centralized control over the information flow from sensory inputs to motor outputs and access to semantic memory for knowledge retrieval. Symbolic architectures can capture high-level cognitive functions and knowledge. Explanation-based, analogical, and inductive learning can be used to derive further know­ ledge. However, they generally lack the means to learn symbolic entities from low-level information and to cope well with large amounts of information and uncertainty. Other challenges concern system scalability and parallel processing abilities. Emergent systems, such as the Local, Error-Driven, and Associative Biologically Realistic Algorithm (LEABRA) architecture,4 are generally inspired by the bottom-up connectionist model of low-level activation signals flowing through a network of processing nodes. These nodes can self-organize and interact in specific ways, modify their internal states, and capture properties of interest. Emergent architectures give the model the context specificity of human performance and enable it to handle information

Accordingly, cognitive architectures require the elaboration and experimental validation of salient features such as knowledge consolidation, scalability, and metacognition to realize true human-like intelligence, thereby contributing back to our understanding of human cognition. With this article, we present the Integrated Neuro-Cognitive Architecture (INCA), which emulates the putative functional aspects of various major brain systems via a learning memory modeling approach. (See the “Developing Cognitive Architectures” sidebar for previous work in this area.) INCA features scalable structural and parameter self-organizing july/august 2011

concurrently and robustly. Most of these architectures (except LEABRA), however, suffer from stability issues when acquiring novel information−a paradigm known as catastrophic interference. Another major issue is the difficulty in deriving the declarative knowledge and higher-level functions crucial to achieve true human-like intelligence. Comparing the relative strengths and limitations of symbolic and emergent architectures shows that they are complementary. By integrating the two, such that each can remedy the other’s deficiencies, we can envisage comprehensive, brain-like architectures covering all levels of processing, from sensory stimuli to high-level cognition. Accordingly, researchers have developed several hybrid architectures in recent years, such as the Connectionist Learning Adaptive Rule Induction Online (Clarion). 5 However, most hybrid architectures still suffer from limited scalability and lack truly flexible self-organizing features, particularly for the automatic construction of memory structures without prior knowledge. Other prominent issues include the efficient integration of multiple learning modalities and long-term knowledge consolidation.

References 1. E.R. Kandel, J.H. Schwartz, and T.M. Jessel, Principles of Neural Science, 4th ed., McGraw-Hill, 2000. 2. W. Duch, R.J. Oentaryo, and M. Pasquier, “Cognitive Architectures: Where Do We Go from Here?” Proc. 1st Artificial General Intelligence Conf., P. Wang, B. Goertzel, and S. Franklin, eds., IOS Press, 2008, pp. 122–136. 3. A. Newell, Unified Theories of Cognition, Harvard Univ. Press, 1990. 4. R. O’Reilly and Y. Munakata, Computational Explorations in Cognitive Neuroscience: Understanding of the Mind by Simulating the Brain, MIT Press, 2000. 5. R. Sun and F. Alexandre, Connectionist Symbolic Integration, Erlbaum, 1997.

mechanisms to form high-level symbolic knowledge from low-level data and knowledge-exploitation mechanisms based on plausible consolidation and inference cycles, respectively.

Integrated Neuro-Cognitive Architecture To address these issues, we designed INCA to be a new hybrid framework that integrates the strengths of symbolic and emergent architectures. 6 (See the sidebar for previous research.) Each INCA module features autonomous structural and para­ meter self-organizing mechanisms utilizing various unsupervised, supervised, and reinforcement learning www.computer.org/intelligent

methods. Unsupervised learning employs clustering, while supervised and reinforcement learning involve mapping the modules’ states to target actions and state-action pairs to a utility function, respectively. In turn, these let us add support for metacognitive (reflexive) functions (such as attention shifting, affect modulation, and executive control) to orchestrate and enhance ongoing processes. INCA’s macro-level behaviors emerge from module interaction in the form of consolidation and inference cycles, epitomizing the human knowledge acquisition and retrieval processes, respectively. This article expands on previous work by 63

Neuro-Cognitive Architectures

Affect modulator (amygdala)

Executive manager (frontal cortex)

Transient memory (hippocampus)

Consolidation

Attention focus

Retrieval

Sensory register (sensory cortex/ thalamus)

Declarative memory (posterior cortex)

Procedural memory (cerebellum/basal ganglia)

Motor register (motor cortex)

Environment Figure 1. Integrated Neuro-Cognitive Architecture (INCA). The main cognitive modules with interconnecting knowledge consolidation and retrieval protocols. The executive manager, declarative memory, and procedural memory collectively constitute the system’s long-term memory.

providing extended specifications of these cycles. 5 The INCA framework consists of several modules: sensory and motor registers, a transient memory, an affect modulator, an executive manager, and distinct declarative and procedural memories (see Figure 1). The last three collectively constitute the system’s long-term memory. INCA also includes several algorithmic protocols to support intermodule communications: attention, consolidation, and retrieval protocols. Apart from the two registers, specified during deployment, all modules are self-organizing memories and thus exhibit functionalities that are learned rather than innate. Although each module maps to a salient brain region, the aim is not to model accurate biological details, which would demand a more complex, less-scalable design and impact system performance. Instead, the modules realize 64

the putative, functional aspects of the brain regions via a neuro-fuzzy system (NFS) modeling approach. NFS is a powerful, hybrid symbolicconnectionist model that exploits a neural network’s learning ability, parallelism, and robustness to induce high-level fuzzy linguistic rules from low-level data.7,8 It facilitates humanlike approximate reasoning by handling uncertainty and imprecision in numerical data while allowing highlevel processing at the semantic and symbolic levels. Architectural Modules

The sensory and motor registers in INCA are the two elementary interfaces for receiving sensory inputs from and executing motor commands in the environment, respectively. They implement preprocessing or postprocessing procedures for simulating the functional (not biologi­ cal) aspects of the reactive, primitive www.computer.org/intelligent

perceptual, and action systems in the brain. Currently, we assume that sensory and motor registers can extract relevant sensory inputs and reliably execute motor commands output, or else delegate those functionalities to other subsystems. The transient memory is a mediator for the formation of the long-term declarative, procedural, and executive memories. It emulates the role of the hippocampus in rapidly encoding perceptual information contributing to our experiences or episodic knowledge.4 To realize this, we developed a fast-learning globalist NFS,7 where the system learning and inference are achieved via global acti­ vation of the entire rule base. We configured the NFS to exhibit a pattern separation trait—that is, distinct events are encoded using separate, nonoverlapping representation. Existing long-term memory content might conversely be used to direct learning or interleave the old know­ ledge with new information acquired in the transient memory. The declarative memory captures general facts about the world or self (the “what”) that are relatively static over time. In contrast to the rapid, globalized learning of transient me­ mory, it performs a slow, localized learning atop a rule neighborhood that reflects the stable storage in the posterior cortex,4 capturing salient (statistical) properties of the world. We built a localist NFS on this principle, 8 where learning and inference use at any time only a small portion of the knowledge base. The model also includes dimensionality- and rulereduction mechanisms to enhance scal­ ability and interpretability. The procedural memory learns and retains knowledge about working skills or procedures (the “how”). It regulates the generation of motor actions, where movements are learned via local error correction on embedded IEEE INTELLIGENT SYSTEMS

synaptic weights, corresponding roughly to the functional role of the cerebellum and basal ganglia regions in the brain.9 This module also provides modulatory (gating) signals to the executive manager, emulating the role of basal ganglia in switching between rapid updating and robust maintenance of information in the frontal cortex.4 These roles can be simulated via the localist NFS, 8 which is now focused on capturing procedural knowledge. INCA’s metacognitive functions are governed by the affect modulator and executive manager. The af­ fect modulator is concerned with motivations and their interactions (the “why”), and its function is tailored within the executive manager. This module provides INCA with a context in which goal setting and reinforcement signals are determined for the other subsystems. Cognitively, we can think of it as a functional analogue of the limbic circuit in the brain, especially the amygdala, which provides the goals to direct system operations and modulate the memoryconsolidation processes.9 A plausible implementation of this module is via the globalist NFS.7 Lastly, the executive manager is used for active monitoring, control, and orchestration of cognitive processes to enhance overall system performance. This role maps functionally to the frontal cortex, 9 such as to change or interrupt ongoing processes in the transient, procedural, and declarative memories; tune external (free) parameters; and so on. Regulation is also achieved via setting reinforcement functions on the basis of motivation signals from the affective modulator. For these purposes, the executive manager includes several submodules for goal setting, parameter tuning, and reinforcement metrics. Each can be implemented juLY/auGuST 2011

algorithmically or via the localist NFS.7 Communication Protocols

The attention­focus protocol in INCA realizes the concept of selective attention in human cognition. The goal is to select the most relevant sensory features or combine existing features into more informative ones, possibly guided by the goal and/or contextual signals from the affect modulator and executive manager. In turn, this allows focused cognitive processing that reduces the other modules’ computational load. Various feature selection and combination methods can be used to realize this role.8

iNCA’s macro-level operations are governed by its consolidation and inference cycles. The retrieval protocol exploits the (associative) knowledge base captured in all INCA modules and is central to its inference cycle. This can take various forms, from aggregating the outputs of the transient and declarative, procedural, and executive memories to achieve a performance surpassing that of individual modules, to retrieving goal and contextual information from the affect modulator and executive manager to modulate operations. Results are ultimately transmitted to the motor register to produce actions or to the sensory register in the next cycle. The consolidation protocol, central to the INCA consolidation cycle, realizes the transfer of knowledge www.computer.org/intelligent

initially stored in the transient memory (which learns fast, but is more plastic) into the long-term storage (which is more resilient, yet learns slowly). Our approach is specifically built on the idea of interleaved learning between the transient hippocampal memory and long-term cortical memory, 4 modulated by the affect signals from the amygdala memory. New information is fi rst kept wholesale in the hippocampus and then continually read into the cortex and interleaved with previously acquired knowledge. This procedure allows optimizing the cortex for the gradual discovery of shared structures (experiences) while enabling the hippocampus to readily absorb new information (exploration) without interfering with existing knowledge. The basic infrastructure supporting these three protocols involves a dual­consolidation network (DCN) that consists of a slow­learning module (SLM) and a fast­learning module (FLM), realized using the slow-learning localist NFS 8 and fast-learning globalist NFS.7 DCN also includes an attention layer running the attention-focus protocol to reduce problem dimensionality and a combination layer aggregating the output signals from SLM and FLM to partly support the retrieval protocol. Meanwhile, the two modules communicate via the consolidation and inference links, which form the primary pathways facilitating the consolidation and retrieval protocols, respectively. Figure 2 shows an example of a DCN architecture comprising two (attended) inputs and a single output.

INCA’s Operational Cycles INCA’s macro-level operations are governed by its consolidation and inference cycles, which can run in parallel and interact through the system’s knowledge base. For illustration 65

Neuro-Cognitive Architectures

Slow-learning module (declarative/procedural/executive) Input layer

Attention layer

LR

LR CL

Input

Output layer

Combination layer

LR AL LR Output

AL

SR

. . .

SR

Affect signal (from affect modulator)

SR SR

: Consolidation link : Inference link

AL

CL

Fast-learning module (transient)

AL: Antecedent label CL: Consequent label SR: Short-term rule LR: Long-term rule

Figure 2. Dual-consolidation network (DCN) in INCA. A novel neuro-fuzzy system comprising fast-learning transient memory and slow-learning declarative, procedural, and executive memory.

purposes, we provide an open-ended scenario of automated driving skill acquisition by an intelligent agent to complement the descriptions of the two cycles. Our earlier work focused on using the INCA procedural memory to learn parking, U-turn, and other driving maneuvers, and more recently included simple tactical decision making.10 In this article, we present several cases requiring higher-level cognitive functions to develop more sophisticated driving skills resembling that of humans. The project thus aims not only at realizing autonomous driving abilities but also understanding and modeling the acquisition of complex cognitive skills. In this scenario, the INCA-based driving agent operates in a virtual 66

traffic world and has innate senses and effectors corresponding to the sensory and motor registers, respectively. The sensory register realizes several primitive sensing functions to extract features of interest from the environment (such as edge detection from visual inputs and a low-pass noise filter for auditory inputs). Meanwhile, the effectors are realized in the motor register that sends out driving control signals to the environment (such as steering, acceleration and brake efforts, and gear shifting). Consolidation Cycle

INCA’s consolidation cycle constitutes the primary mechanism for acquisition, transfer, and synthesis of the entire knowledge base, carried out internally at recurring times www.computer.org/intelligent

(periodical or event-based) whenever knowledge acquisition and updates are necessary. Our general assumption is that a task consists of various experiences that can be divided into independent data chunks occurring at different time periods (episodes). Figure 3a summarizes the major steps of the consolidation cycle. Initialization. The attention focus protocol is first executed to select or combine relevant sensory features, possibly directed by the executive manager—for example, learning to attend to the left mirror stimulus before switching to the left lane. Based on these attended features, the transient memory performs a rapid online learning procedure to adapt both its structure and parameters IEEE INTELLIGENT SYSTEMS

dynamically for every new data pattern received. In addition, the transient memory can retrieve patterns reinstated from the long-term memory (termed pseudo­ patterns) that reflect part of the existing knowledge therein. Each pseudopattern is derived based on the parameters of a rule randomly selected in long-term memory. When the transient memory needs to expand beyond capacity, a deletion procedure (such as the least recently used rule) is carried out to keep the memory bounded (as is the hippocampus). In our example, this concerns rapid encoding in the transient memory of recent driving traces and/or some previous experiences reinstated from the long-term driving rule base. Structural reorganization. Associative learning uses pseudopatterns reinstated from the current transient memory representation to craft the coarse-rule-base structure in the long-term memory. It creates many (linguistic) label neurons, removes those of low significance, and allocates space for the (fuzzy) rules based on the surviving neurons.8 This process emulates the initial development stage of human cortical memories, where coarse structure is first laid out by an activity-independent process.9 Active neurons then stabilize via the uptake of trophic factors, whereas inactive neurons eventually die. In the driving example, fuzzy labels are formed such as (Mild, Steep) for road angle or (Far, Near) for distance to obstacle, and the corresponding rules are generated—for example, “If road angle is steep and distance to obstacle is near then decision is slow down.” The rule-reduction process then discards antecedent links that do not contribute enough and prunes duplicate rules.8 A highly intuitive and concise set of rules can july/august 2011

Initialization

Input/output patterns

• Transient memory • Executive manager

Structural reorganization • Declarative memory • Procedural memory • Executive manager

Yes

Interleaved learning • Declarative memory • Procedural memory • Transient memory • Executive manager • Affect modulator Yes

Knowledge base

No

Optimize more?

(a) Perception • Sensory register

Input patterns

Reflexive learning • Executive manager • Procedural memory • Affect modulator

No

Another episode?

Attention focus • Sensory register • Executive manager

• Executive manager • Affect modulator

Reflexive control

Action selection

Association retrieval

• Procedural memory • Motor register

• Transient memory • Declarative memory

Yes Decision outputs

No

More inputs?

(b)

Figure 3. The two operational cycles in INCA. (a) The consolidation cycle for knowledge acquisition, reduction, and transfer to long-term memory, and (b) the inference cycle for attended perception, knowledge retrieval, and decision making. The modules involved at each step are listed in bullets.

thus be attained. For example, the driving rule is simplified to “If distance to obstacle is near then decision is slow down” because the angle becomes insignificant when the distance is critical. Further reorganization could involve chaining rules and building hierarchies of sequential behaviors. 3 These form the primary mechanisms for prediction and planning, on which higher-level cognition is built, such as in our scenario to realize driving anticipation and navigation. Interleaved learning. This phase involves a rehearsal of the pseudopatterns generated from the transient memory to fine tune the weight parameters of the long-term memory. The pseudo­ patterns are derived in the same way as that from long-term memory in the initialization phase. Feeding together the new and old information into the long-term memory leads to an interleaved learning www.computer.org/intelligent

regimen that incorporates new infor­ mation gradually while refreshing exis­ ting knowledge so the system will not forget it. This procedure conforms to the idea that catastrophic interference can be minimized by interleaving old patterns as new patterns are learned,4 and the hippocampus is there to provide initial storage of patterns in a way that avoids interference. In our example, interleaved learning involves reinstatements of the recent and old driving patterns stored and using them together to tune longterm semantic driving rules. Such a procedure is particularly useful to gracefully handle novel traffic situations that are inconsistent with current, long-term driving expertise. Reflexive learning. This phase serves

to evaluate the performance of INCA modules and optimize their operations with respect to the task at hand. It constitutes INCA’s core meta­ cognitive mechanism, involving the 67

Neuro-CogNitive ArChiteCtures

executive manager, affect modulator, and procedural memory, to allow selfawareness (the agent knows what it is doing) and self-improvement (it knows what to do) over time. This can be realized via a search in the space of the modules’ external (free) parameters, emulating the role of neuromodulators (such as dopamine, and acetylcholine) within the frontal cortex, basal ganglia, and amygdala circuits in providing global regulatory signals for human learning.4,9 Such a meta-optimization approach is independent of the module’s internal confi guration, architecture, or learning method, and it allows evaluating and switching between different (sub)modules. An example would be tuning the driving agent’s free parameters to improve parking performance over time, from the early, erratic execution phase to the expert stage. Inference Cycle

INCA’s inference cycle realizes a serial flow of reasoning processes, starting from perception and ending with action, performed continuously atop the current system’s knowledge base. It offers a plausible description of the human knowledge exploitation processes and shares some similarities with the nature of system engineering life cycles. Figure 3b summarizes the phases of the inference cycle. The perception phase is concerned with primitive information processing based on senses. Specifically, input stimuli from the environment are fi rst kept and encoded in the INCA sensory register. A basic computational analysis is performed on stimuli containing visual, auditory, and other forms of perceptual information to extract the features of interest to the current task. For instance, the sensory register for our driving agent performs task-specific edge detection 68

on its visual inputs to extract the road angle as well as active vision-based depth estimation to compute the distance to the nearest objects. In the attention­focus phase, the features extracted in the perception phase are fi ltered via the attentionfocus protocol, possibly directed by the executive manager and affect modulator. This process can be viewed as a form of competition for attention based on how informative a feature subset or combination compares to others. The attended input features are then broadcast as cues for knowledge retrieval and further processing in the remaining INCA modules.

the iNCA executive manager and affect modulator work jointly to orchestrate the operations of the transient, declarative, and procedural memory modules. One instance of the attention-focus procedure in the driving scenario is to decide whether to attend to the visual inputs from the left, right, and rear mirrors before initiating a lane change. During the refl exive­control phase, the INCA executive manager and affect modulator work jointly to orchestrate the operations of the transient, declarative, and procedural memory modules. This might be in the form of setting a goal context for the three modules, rerouting their operations, altering their inference mode, www.computer.org/intelligent

or setting their free parameters. In effect, these enable the system to become aware of and regulate its present state and performance. In our driving example, these concern the tactical level (such as to determine which operational driving module can best handle a maneuver or traffic situation or to set relevant goal information). In the association­retrieval phase, the associative knowledge stored in the transient and declarative memories is invoked and aggregated (using the retrieval protocol) based on currently attended stimuli and signals from the executive manager and affect modulator. This may be achieved at different complexity levels either through reactive inference, which generates decisions by directly invoking the association rules, or deliberative inference, which plans decision or action sequences to achieve some goal. An example of the former is the fi ring of lane switching rules to overtake a slow vehicle ahead, and an example of the latter is the composite decision sequence necessary to exit the highway. Lastly, during the action­selection phase, the aggregated outputs from the transient and declarative memories are propagated to the motor register for the execution of motor commands in the environment or back to the sensory register for the execution of internal actions in the next cycle. They may be transmitted directly or indirectly via the procedural memory. For example, if the aggregated decision output suggests a lane change before negotiating a road exit and sensors indicate that it is vacant, the procedural memory might initiate a steering command, which gets forwarded to the motor register to adjust the position of the car, or to the sensory register to provide proprioceptive information. IEEE INTELLIGENT SYSTEMS

Episode 2

Episode 1 (Ap,Bp)

1

(Ap,Cp)

(Ap′,Bp′)

3

Recall mean square error (episode 1)

offline manner. For simInterleavedplicity, we do not supLearning Example ply an external pattern in To illustrate the working FLM SLM FLM SLM this step. Based on the principles of DCN and pseudopatterns, the SLM the INCA consolidation Initialization Initialization initializes its structure cycle and how they re(Ap′,Bp′) or (also forming 20 rules) solve catastrophic inter(Ap′,Bp′) (Ap′,Cp′) 2 4 and then performs F epference, we provide an ochs of parameter tuning example of a sequential(via the least-mean-square learning task, which is FLM SLM FLM SLM method 7) to gradually both a fundamental cognitive skill and a classic grasp the FLM representa• Structural reorganization example of forgetting in tion of the A–B patterns. • Structural reorganization • Interleaved learning • Interleaved learning neural networks.4 The maximum number the maximum number The task consists of of rehearsal epochs F is two episodes receiving the Figure 4. Workflow of the dual consolidation network (DCN) for 1,000 in this case. same stimuli Ap but trig- a sequential list learning task. Two-episode example of learning association A–B followed by A–C. During episode two gering different response (step three), a set of P patterns Bp and Cp, repseudopatterns (A′p, B′p) spectively (where p = 4.5 1, …, 20). Items Ap, Bp, (where P << N) are genP=0 P=1 erated from the SLM in and Cp are represented 4.0 P=2 the same way as that from by binary-valued vectors, P=3 3.5 FLM in episode one, and each comprising 5 bits P=4 it is then fed into FLM towith values 0 or 1 equally 3.0 gether with each new exlikely. The DCN is tasked 2.5 ternal pattern (A p, C p). to sequentially learn the set of A–B associations Whenever a new rule needs 2.0 (episode one) and then to be added in the FLM to A– C associations (epicapture an incoming pat1.5 sode two). For simplicity, tern and its current size is 1.0 the consolidation cycle we at maximum capacity, the describe here does not least recently used rule is 0.5 include the reflexivedeleted. learning phase. In step four, N pseu0 Figure 4 gives the DCN dopatterns are gener0 100 200 300 400 500 600 700 800 900 1000 workflow for this task. ated from the FLM, now Rehearsal epoch (episode 2) Both the SLM and FLM containing (A′p, B′p) and are initially empty. In epi­ Figure 5. DCN recall error trace for A–B association after (A′p, C ′p) patterns resode one, the FLM first learning A–C in episode two. Without interleaving (P = 0) the flecting the A–B and carries out rapid online system completely forgets the first association, whereas with A–C associations, to tune learn­ing to capture each interleaving it remembers both (albeit partially). the SLM’s structure and external pattern (Ap, Bp), parameters. These correas per the consolidation cycle’s ini- FLM (ideally) learns episodes one spond to the structural reorganization and interleaved-learning phases tialization phase. No attention- by one. In step two, the FLM repeat- in the INCA consolidation cycle, focus step is assumed, and the FLM’s maximum capacity is set to the num- edly generates N pseudopatterns respectively. We conducted simulations to study ber of patterns N in one episode (A′p, B′p) from the centroid coordi(20 rules), which emerges natu- nates of randomly selected rules and the effects of the number of pseurally from the consideration that the uses them to train the SLM in an dopatterns P (transmitted from SLM july/august 2011

www.computer.org/intelligent

69

Neuro-Cognitive Architectures

The Authors Richard J. Oentaryo is a researcher at Nanyang Technological University, Singapore. His research interests include brain-inspired cognitive architectures, meta-learning, and hybrid intelligent systems. Oentaryo has a PhD in computer science from Nanyang Technological University. He is a student member of IEEE and the IEEE Computational Intelligence Society. Contact him at [email protected]. Michel Pasquier is an associate professor at the American University of Sharjah, United

Arab Emirates. This work was performed while he was director of the Centre for Computational Intelligence, Nanyang Technological University, Singapore. His research interests include cognitive systems, adaptation and learning, and nature-inspired systems. Pasquier has a PhD in computer science from the National Polytechnic Institute, Grenoble, France. He is a member of IEEE and the IEEE Computational Intelligence Society. Contact him at [email protected].

to FLM) on the recall performance of the old A–B patterns while learning A–C patterns in episode two. Figure 5 shows the recall trace across different rehearsal epochs measured in terms of mean-square error, with P varied from 0 to 4. Catastrophic interference is evident in the case of P = 0, which is when no interleaving occurs, similar to that of a classic connectionist system with strict sequential learning. The benefit of interleaving is already clear with P = 1, with a low error rate maintained after rehearsal. Further increases in P (from 2 to 4) yield even lower interference levels, with the A–B association kept nearly intact. These results demonstrate the complementary roles of SLM and FLM in the DNM and, more generally the consolidation cycle, in suppressing catastrophic interference, in sharp contrast with the sequential learning paradigm in other systems. This procedure can be applied to more (greater than two) episodes, possibly involving multiple unrelated tasks and experiences.

architectures that will guide their realization. To this end, the proposed INCA framework, when further developed, might play a key role toward providing scalable and flexible selforganizing mechanisms, formulating the consolidation and inference cycles that govern robust knowledge acquisition and exploitation, and building support for metacognitive abilities and self-improvement. Further results of this research (available as a Web extra at http:// doi.ieeecomputersociety.org/10.1109/ MIS.2011.60) include a detailed study demonstrating INCA’s scalability and generalization traits, where a biomedical domain with 927 features and several thousand data points11 was reduced to 19 rules providing an 88-percent prediction accuracy on par with the best methods available. Investigations are now underway to experiment with reflexive learning and control algorithms, thereby enabling INCA to acquire and exploit skills and concepts autonomously, from sensory-motor tasks all the way to the highest levels of cognition.

T

References

he ever-increasing complexity of today’s systems and devices emphasize the need for natural communication and human-like learn­ ing and cognitive abilities, hence for a more general machine intelligence, and for developing cognitive

70

1. D. Michie and R. Johnston, The Creative Computer, Viking, 1984, p. 214. 2. A. Newell, Unified Theories of Cogni­ tion, Harvard Univ. Press, 1990. 3. R. Sun and F. Alexandre, Connectionist Symbolic Integration, Erlbaum, 1997. www.computer.org/intelligent

4. R. O’Reilly and Y. Munakata, Com­ putational Explorations in Cognitive Neuroscience: Understanding of the Mind by Simulating the Brain, MIT Press, 2000. 5. W. Duch, R.J. Oentaryo, and M. Pasquier, “Cognitive Architectures: Where Do We Go from Here?” Proc. 1st Artificial General Intelligence Conf., P. Wang, B. Goertzel, and S. Franklin, eds., IOS Press, 2008, pp. 122–136. 6. R.J. Oentaryo and M. Pasquier, “Towards a Novel Integrated NeuroCognitive Architecture (INCA),” Proc. IEEE Int’l Joint Conf. Neural Networks, IEEE CS Press, 2008, pp. 1902–1909. 7. R.J. Oentaryo, M. Pasquier, and C. Quek, “GenSoFNN-Yager: A Novel Brain-Inspired Generic SelfOrganizing Neuro-Fuzzy System Realizing Yager Inference,” Expert Systems with Applications, vol. 35, no. 4, pp. 1825–1840, 2008. 8. R.J. Oentaryo and M. Pasquier, “A Reduced Rule-Based Localist Network for Data Comprehension,” Proc. IEEE Int’l Joint Conf. Neural Networks, IEEE CS Press, 2008, pp. 2660–2667. 9. E.R. Kandel, J.H. Schwartz, and T.M. Jessel, Principles of Neural Science, 4th ed., McGraw-Hill, 2000. 10. M. Pasquier and R. J. Oentaryo, “Learning to Drive the Human Way: A Step Towards Intelligent Vehicles,” Int’l J. Vehicle Autonomous Systems, vol. 6, nos. 1–2, 2008, pp. 24–47. 11. A.G. Pedersen and H. Nielsen, “Neural Network Prediction of Translation Initiation Sites in Eukaryotes: Perspectives for EST and Genome Analysis,” Proc. Int’l Conf. Intelligent Systems for Molecular Biology, AAAI Press, 1997, pp. 226–233.

Selected CS articles and columns are also available for free at http://ComputingNow.computer.org. IEEE INTELLIGENT SYSTEMS

Cognitive Architecture

Learning Adaptive Rule Induction Online (Clarion).5. However, most ... that reflects the stable storage in the posterior ..... “Learning to Drive the Human Way: A.

2MB Sizes 3 Downloads 182 Views

Recommend Documents

A cognitive architecture for emergency response
user's attention as the server and tasks as jobs (see [10] for details). As jobs arrive ..... ACM Transactions on. Computer-Human Interaction, 13(1):37–70, March.

Cognitive discoordination 1 Cognitive discoordination ...
Brain Research Reviews, 41, 57-78. Leiser, D., & Bonshtein, U. (in press). Theory of mind in schizophrenia: Damaged module or deficit in cognitive coordination ...

III. The Meta-cognitive view cognitive view - Sites
May 26, 2011 - model builders. B. Reading text: devoid of meaning by itself. C. Reading: An interactive process. 3. Mohammed Pazhouhesh. III.

The Production of Cognitive and Non- Cognitive ... - Yao Amber Li
"The Colonial Origins of Comparative Development: An Empirical Investigation." American Economic. Review 91.5 (2001): 1369-1401. [2] Autor, David H., Frank Levy and Richard J. Murname, 2003. “The Skill Content of Recent Technological Change: An Emp

The Production of Cognitive and Non- Cognitive ...
measure the quantity of education (i.e. years of schooling or expenditures per pupil), it is diffi cult to ... low international scores despite having one of the highest levels of per-capita educational spending ... evaluation and rewards for their t

Cognitive Bubbles - macroeconomics.tu-berlin.de
Nov 10, 2015 - software, and Pablo Lopez-Aguilar for helping with its implementation. ..... were incentivized to nudge subjects to give their best guess of .... in the race game,” Journal of Economic Behavior & Organization, 75(2), 144–155.

Cognitive Bubbles
Dec 10, 2017 - different sets of instructions for the same game, and by introducing hints, they show that subjects do not deviate from .... The resulting distribution of beliefs provides an estimate of what subjects think about ... independently of t

THE COGNITIVE INTERVIEW The Cognitive Interview ...
to make recommendations for policy and practice based on the 1999 ... line searches, researchers in the field were contacted via obtaining email lists from ... (enhanced CI vs. structured interview), retention interval (four hours vs. six weeks), ...

The Production of Cognitive and Non-cognitive Human ...
We accommodate incen- tives in our model by having heterogeneous workers make optimal occupational choices given their own comparative advantages in ...

Cognitive Architectures
or categorization, are accomplished on a faster time scale in a parallel way, without ... scale knowledge bases, bootstraping on the resources from the Internet. ..... The SNePS Rational Engine controls plans and sequences of actions using.

Cognitive dissonance
influence the believability and performances of mediated experiences. Perceptual conflicts are ... truth and cause wrong decisions. This is true for perceptual conflicts too, where mid-way solutions can be described that do not correspond to the feat

Cognitive Bubbles - macroeconomics.tu-berlin.de
Nov 10, 2015 - archives/2014/11/vernon_smith_on_2.html. ... second dates proposed in the email varied between one and five weeks after the initial ...... the correct number to add up to 60 if the sum is above 49, otherwise the computer plays.

Architecture Architecture docx.pdf
Optional Architecture Courses. Business Information Management I (BIM). Credit: 1.0. Grade: 9-12. Counts as Technology graduation credit. Interior Design.

Cognitive Bubbles - macroeconomics.tu-berlin.de
Mar 31, 2016 - This setup has two advantages. First ..... 4 Results. We begin the analysis of our data by looking at the baseline treatments in section 4.1, and the ... ments look very much alike, despite the big payoff differences. Indeed, a ...

Cognitive Modeling
mators are given the right tools for the job. To this end, we ... umbrella of this description, but the distinguishing features of. CML are .... the distinguished function do, so that s = do(a, s). The possi- ..... In the territorial T-Rex animation

Cognitive Walkthrough Forms
University of Colorado at Boulder. Department of Computer Science ... Alto, California designed and completed a study to compare a variety of usability ..... performed; where each "best way" corresponds to one sequence of user actions, eg. a ...

Cognitive Test Reviews
Excluded from norm sample if diagnosis of cognitive impairment, loss of consciousness, CVA, epilepsy, CNS infection, CNS disease, or head injury; uncorrected vision/hearing loss; non-fluent in. English; diagnosis/history of alcohol/drug dependence; m

Cognitive Mobile Homes
6While “foundationalism” is used in too many different ways to speak of a “standard” use, my formulation is similar to what many .... While it is easy to see how one might appeal to empirical linguistic data to adjudicate debates about the ..