<?xml version="1.0"?>
<document xmlns="http://cnx.rice.edu/cnxml" xmlns:md="http://cnx.rice.edu/mdml/0.4" xmlns:bib="http://bibtexml.sf.net/" xmlns:q="http://cnx.rice.edu/qml/1.0" id="id3995413" module-id="m12345" cnxml-version="0.6">
  <title>Neurocognitive Informatics Manifesto</title>
  <metadata>
  <!-- WARNING! The 'metadata' section is read only. Do not edit below.
       Changes to the metadata section in the source will not be saved. -->
  <md:content-id>new</md:content-id>
  <md:title>Neurocognitive Informatics Manifesto</md:title>
  <md:version>**new**</md:version>
  <md:created>2009/07/22 07:42:45.010 GMT-5</md:created>
  <md:revised>2009/07/22 07:44:39.536 GMT-5</md:revised>
  <md:authorlist>
    <md:author id="WDuch">
        <md:firstname>Wlodzislaw</md:firstname>
        <md:surname>Duch</md:surname>
        <md:fullname>Wlodzislaw Duch</md:fullname>
        <md:email>wduch@is.umk.pl</md:email>
    </md:author>
  </md:authorlist>
  <md:maintainerlist>
    <md:maintainer id="WDuch">
        <md:firstname>Wlodzislaw</md:firstname>
        <md:surname>Duch</md:surname>
        <md:fullname>Wlodzislaw Duch</md:fullname>
        <md:email>wduch@is.umk.pl</md:email>
    </md:maintainer>
  </md:maintainerlist>
  <md:license href="http://creativecommons.org/licenses/by/3.0/"/>
  <md:licensorlist>
    <md:licensor id="WDuch">
        <md:firstname>Wlodzislaw</md:firstname>
        <md:surname>Duch</md:surname>
        <md:fullname>Wlodzislaw Duch</md:fullname>
        <md:email>wduch@is.umk.pl</md:email>
    </md:licensor>
  </md:licensorlist>
  <md:keywordlist>
    <md:keyword>Cognitive systems,</md:keyword>
    <md:keyword>Medical ontologies;</md:keyword>
    <md:keyword>Natural language processing;</md:keyword>
    <md:keyword>Neural networks,</md:keyword>
    <md:keyword>Neurocognitive informatics,</md:keyword>
    <md:keyword>Semantic networks;</md:keyword>
    <md:keyword>Spreading activation networks;</md:keyword>
    <md:keyword>Vector models in NLP</md:keyword>
  </md:keywordlist>
  <md:subjectlist>
    <md:subject>Science and Technology</md:subject>
  </md:subjectlist>
  <md:abstract>Informatics studies all aspects of the structure of natural and artificial information systems. Theoretical and abstract approaches to information have made great advances, but human information processing is still unmatched in many areas, including information management, representation and understanding. Neurocognitive informatics is a new, emerging field that should help to improve the matching of artificial and natural systems, and inspire better computational algorithms to solve problems that are still beyond the reach of machines. In this position paper examples of neurocognitive inspirations and promising directions in this area are given.</md:abstract>
  <md:language>en</md:language>
  <!-- WARNING! The 'metadata' section is read only. Do not edit above.
       Changes to the metadata section in the source will not be saved. -->
</metadata>

<content>
    <para id="id1166113551414">
      <footnote id="id7511391"/>
    </para>
    <para id="id1594110">
      <emphasis effect="bold">Neurocognitive Informatics Manifesto</emphasis>
    </para>
    <para id="id1862132"><emphasis effect="bold"/>Włodzisław Duch</para>
    <para id="id1166118812207">Department of Informatics, Nicolaus Copernicus University,</para>
    <para id="id1166113792765">Grudziądzka 5, Toruń, Poland.</para>
    <para id="id1166113799541">Google: W. Duch</para>
    <para id="id7987895">Abstract</para>
    <para id="id2935880">Informatics studies all aspects of the structure of natural and artificial information systems. Theoretical and abstract approaches to information have made great advances, but human information processing is still unmatched in many areas, including information management, representation and understanding. Neurocognitive informatics is a new, emerging field that should help to improve the matching of artificial and natural systems, and inspire better computational algorithms to solve problems that are still beyond the reach of machines. In this position paper examples of neurocognitive inspirations and promising directions in this area are given. </para>
    <para id="id2923314"><emphasis effect="italics">Keywords</emphasis>: Natural language processing; Semantic networks; Spreading activation networks; Medical ontologies; vector models in NLP</para>
    <section id="id5179642">
      <title>Introduction</title>
      <para id="id3086973">Technology is developing at an unprecedented pace, from simple systems that were easy to control to handheld devices, such as personal digital assistant (PDA) computers or smart telephones with multimedia capabilities, devices with so many options that rarely more than a tiny fraction are ever used even by quite intelligent people. Further miniaturization of integrated circuits will only increase complexity of the environment, exceeding capability of cognitive information processing by most humans. In other areas situation is not better: only a small fraction of all functions is used in a typical word processor or other software, equipment used for medical diagnosis is capable of many sophisticated measurements, only a fraction of which is understood and relied on by medical doctors. Human brain is highly adaptable due to the neuroplasticity , therefore the speed of reaction and the ability to deal with complex devices is increasing, but very few people can become experts in more than one area. </para>
      <para id="id1166113341546">The increasing dependence on the sophisticated technology that should help with the complexities of modern life poses requires revolutionary changes in organization of information. The great challenge for informatics is thus in creation of interfaces that will make technology more human, easier to use, adjusted to our ways of information processing. Humanized interfaces imply the ability of software to anticipate human reactions, understand questions and errors that humans are prone to make, ask questions to specify precisely the request, in general adjust itself to the human-like ways of thinking. These problems are only partially addressed by the current branches of informatics, although there have been some attempts in the past to understand the ways our brains are processing information and to adjust are methods accordingly. Already H. Spencer in “Principles of Psychology'” published in 1855 made a conjecture that all intelligence may be interpreted in terms of successful associations between psychological states driven by the strength of connections between the internal states (see the early history of connectionism in ). Yet to this day we do not have machines that would help us to find interesting associations, coupling to and enhancing our natural abilities. The methods that are currently used to represent knowledge  are not as flexible as natural language and the imagery that brains are capable of. Only human brains are capable of using language, and bringing the existing Natural Language Processing methods closer to what human brains do reacting to and understanding concepts expressed by language structures is ultimately the only certain way to human-level competence in this area. While formal and statistical models of language may capture some important features it is unlikely that the texts itself contain sufficient information to train a disembodied system to fully understand the meaning. </para>
      <para id="id1166113537512">There is general agreement among philosophers of language and neuroscientists who work on language  that the meaning of basic concepts is grounded in the embodiment, the ability to perceive and act. The basic function of the language in the brain is to activate perception/action networks. This trend started in robotics with R. Brooks Cog project manifesto , in philosophy with F. Varela, E. Thompson and E. Rosch 1991 book , and in linguistics it has been represented by G. Lakoff and others . Problems with traditional approach to natural language led some cognitive systems experts to base their hope on developmental approach, hoping that robots acting in the world are going to develop some proto-language exchanging messages . Such efforts are not new, already in 1998 Hutchens and Alder created a chatterbot called MegaHAL  based on stochastic language model that could be trained by interaction with people, but after 10 years this chatterbot is still on the level of using 2-3 word combinations that can be interpreted as capturing some statistical relations from the conversations with humans. There has been some interesting progress in modeling relations between word use, perceptions and actions, relating words to their real referents and helping to define word meanings in specific contexts. These models are still very much restricted to a few areas, such as spatial relations, or color and shape naming, but can be used as an inspiration for simpler representations. Perceptual properties of objects are associated with possible actions that involve these objects, including motor affordances for manipulation. A ball has a round shape, color, size, but also may be kicked or bounced. All this information should be included in the representation of the “ball” concept. </para>
      <para id="id4854649">Developmental approach, as well as fully embodied approaches may provide useful inspirations but are not directly applicable to large scale practical NLP applications. The challenge is to identify the processes and find approximations to what brains are doing and use them to correct and enhance the existing vector-based approaches. This is the foundation for more detailed understanding of higher cognitive processes related to thinking and problem solving. As Feldman writes  “thought is structured neural activity”, and “language is inextricable from thought and experience”. Formal reasoning and linguistic methods are quite useful but quite different from information processing that our brains do. </para>
      <para id="id1166118548628">The goal of neurocognitive informatics is to draw inspirations from recent advances in neuroscience, create practical algorithms, especially in areas where traditional approaches fail, and increase our understanding of higher cognitive processes by providing simple models that explain how our minds work. In “The Cognitive Neurosciences” book  that largely defined the whole field of research, Gazzaniga wrote: “The future of the field, however, is in working toward a science that truly relates brain and cognition in a mechanistic way.” Various approaches to do that will be considered here. In the next chapter relation of neurocognitive informatics to other, existing fields is outlined. The third chapter is mainly focused on language and associative processes. In chapter four some results inspired by the neurocognitive approach to language are presented. The final chapter presents some conclusions and visions of the future of this field. </para>
    </section>
    <section id="id1166118657400">
      <title>Neurocognitive informatics and related fields</title>
      <para id="id4329368">Inspiration from nature have of course always been used in science, including computer science. Computational intelligence emerged from interactions of several research communities with overlapping interests, inspired by observations of natural information processing . The three main branches of computational intelligence include evolutionary computing, neural networks and fuzzy logic. Evolutionary processes that structured sensory organs, brains and intelligence are used as an inspiration for various genetic optimization algorithms . Later also other biologically-inspired optimization approaches, such as ant, swarm, and immunological system algorithms, have been introduced -. All these biologically inspired algorithms are used for similar applications, although the time-scales of evolutionary, behavioral and immunological processes are very different, and the type of intelligence they are related to is also quite different.</para>
      <para id="id7653753">The field of neural networks is quite broad, but mostly inspired by models of single neurons that perform basic computations, models of perception and sensorimotor reactions (primary sensory and motor cortices), a threshold-based perceptron, or spiking neuron being the main computational element. Artificial neural networks (ANNs) are usually based on drastically simplified models of biophysical neurons. Initially logical or graded response (sigmoidal) neurons were almost exclusively used in ANN models  applied for classification, approximation or associative memories. Multi-layer perceptrons (MLP networks) are still the most successful ANN type of models. Vector quantization approaches used for clusterization and self-organization have also been included  to account for unsupervised learning and model development of topographical maps. These model use prototypes and similarity functions that could be weighted in a non-linear way, usually using Gaussian factors, to account for non-linear nature of neural interactions. This led to introduction of basis set expansion techniques, used in approximation theory  and quite common in pattern recognition, with computation flow that may be presented in a graphical form similar to the MLP networks, and thus became “a neural network” , with radial basis function (RBF) networks  becoming a major alternative to multilayer perceptron networks (MLPs). Although ANNs in principle could approximate any function there are some relatively simple problems that such networks are not able to solve, requiring exponentially large number of examples , for example the topological invariants of patterns, the problem of connectedness (determining if the pattern is connected or disconnected). Such problems can only be solved with a different type of neurons that include at least one phase-sensitive parameter , or with spiking neurons . Although feedforward artificial neural networks found wide applications in knowledge extraction from data , simulating human perception and recognition of objects and rules, it has not been used much for knowledge representation or reasoning based on such perceptions. </para>
      <para id="id1166118586904">Computational cognitive neuroscience is mostly focused on biophysical models of neurons, neurodynamics and analysis of various brain signals. Good models of spiking neurons have been developed , and although mathematical characterization of their power has been described  their practical applications are still limited. The goal is to learn about the detailed mechanisms of biological systems, rather than try to abstract essential principles to make useful algorithms. Neuroinformatics is a new branch of science that is focused mostly on support of brain research, databases and tools for analysis of brain imaging and electrophysiological recordings. Neural ideas are also used in affective computing, trying to understand and express emotional content of speech and behavioral acts. Human-Computer Interaction (HCI) is a field that is mostly focused on modeling the cognitive limitations of human information processing. Neural models of higher cognitive functions are still very limited to small-scale experiments  and will certainly be quite complex. Neurocognitive informatics is aimed at abstracting the principles and creating large scale applications, therefore the overlap here is rather limited. </para>
      <para id="id2807440">Fuzzy set theory has its inspirations in extensions of multivalued logics, but fuzzy logic and fuzzy rules may also be related to perception, object recognition, prototype-based categorization  and associative reasoning. These inspirations are more on psychological than neural level. Sets of fuzzy rules have a natural graphical representation , and are deeply connected to neural networks and uncertainty in measurements . In particular if separable functions are used in neural network nodes their outputs may be treated as linear aggregation of fuzzy decisions based on product norms . Fuzzy rules organized in a network form may also be tuned by adaptive techniques used in neural networks, therefore they are called neurofuzzy systems . Neural networks may also use more sophisticated transfer functions, combining weighted activations of their inputs with distance-based activations . In such a case each node does not model a single neuron but a larger group of neurons, perhaps a microcircuit, that performs more complicated function. Connectionist modeling in psychology  goes much further, assuming that network nodes represent whole concepts, and their connections represent explicit relations between these concepts. Each node in fact corresponds to many states of network subconfigurations that code concepts, although exact relations of connectionist models to neural processes have not yet been elucidated. Such networks may still incorporate the competitive learning “winner-takes-most” mechanisms. Semantic networks  simplify this further, simply linking the concepts with connections that designate relations of different types. Connectionist and semantic networks are examples of general graphical models , with other examples including Bayesian belief networks and abstract network models for parallel distributed processing. </para>
      <para id="id4046299">Concept nodes, representing information derived from perception or from some measurements, are not too precise and thus may be represented in terms of fuzzy membership functions. Fuzzy models may be formally derived as generalization of multivalued logics, but the field has also rather obvious cognitive inspirations related to models of intelligent behavior at a higher, psychological level, rather than elementary neural level. Some efforts have been made to use fuzzy ideas in computing with words  and computing with perceptions , but these efforts are restricted to understanding the meaning of a small subset of linguistic phenomena and do not address NLP problems, or other cognitive systems problems. Granular computing becomes fashionable, defining “granules of information” at various level, and although on the mathematical side most algorithms used in this field can be related to clusterization one can imagine distant connections of information granules with activations of different brain structures. Fuzzy and neural systems are at the core of computational intelligence.</para>
      <para id="id1638165">Computational intelligence community is also interested in statistical methods that grew up from research on perceptrons. Better understanding of mathematical foundations of neural techniques resulted in popular statistical pattern recognition models, such as the Support Vector Machines (SVMs)  for supervised learning, or Independent Component Analysis  and similar techniques for unsupervised learning. Probabilistic, Bayesian models, maximum likelihood and other approaches are being added to the bag of Computational Intelligence tools, but very little is being done with these methods in respect to knowledge representation, language or reasoning. </para>
      <para id="id1166112546337">To cover all phenomena related to intelligence in a computational framework representation and reasoning based on complex knowledge structures is needed. Artificial intelligence (AI) has focused on symbolic knowledge representation, addressing problems that require reasoning, planning, understanding of language. These high-level cognition problems require quite different approach than problems that Computational Intelligence is addressing. Reasoning using linguistic concepts requires knowledge representation of diverse types of information that may be combined in an infinite number of ways. Logical, rule-based or frame-based representation schemes are quite useful but in many respect rather limited, inadequate to model associative processes that are always active in the brain. A conceptual blending theory of cognition  is based on assumption that various features and objects may be subconsciously “blended” facilitating creative thinking. This requires associative combinations of features of different representations. Although it is not clear how this process could proceed in the brain the theory has been partially implemented at the symbolic AI level . It is certainly an important source of inspiration for neural models of creative processes and higher cognitive functions. Making inferences from partial observations and integrating them requires systematic search techniques and may draw inspirations from decision-making processes in which prefrontal cortex of the brain is involved. </para>
      <para id="id5716896">One of the big challenges facing CI and AI communities is the integration of the algorithms that serve for modeling of both low and high-level cognitive processes. Autonomous robots need to reason using information based on perceptions, coming from sensors, categorized into information granules that are easier to handle. Although initial steps towards such integration have been made long time ago  many textbooks on artificial intelligence discuss only symbolic machine learning methods  that generate logical rules. Learning the rules directly from signals and measurements is certainly necessary in cognitive robotics and there are good algorithms to achieve this . Recent integration of two major cognitive architectures, the neural Leabra  and the symbolic ACT-R, into one SAL architecture , showed that symbolic AI approaches may be understood as rough approximation to the brain functions. Such hybrid models are able to account for quite complex behavior and higher mental functions better than purely neural models.</para>
      <para id="id1166113353371">Arguably the biggest challenges lie in the language domain. Language is the natural form of communication, both between people and to some degree within the human brain that strives for narrative understanding of its own history. Adding verbal comments to complex internals states helps to recall them, understand and reason about them. The ideal information management system should respond to queries in natural language. A natural “talking-head” interface is relatively easy to add  to chatterbots, but linguistic abilities of such programs are very limited. We shall focus here on neurocognitive inspirations in the language domain, although the goals of neurocognitive informatics are much wider. Providing simple models of higher cognitive functions should help to understand these functions better, and simplify them to create practical algorithms for large-scale applications. These functions include knowledge representation, different forms of memory (recognition, semantic, episodic, procedural, long and short-term, and working memory), connections of language to perception, thinking, problem solving, planning, attention, behavioral control, imagery, creativity and consciousness. </para>
      <para id="id6768485">Although many details of processes in the brain are not known what we have learned so far is already a great source of inspiration. Intuition, insight, imagery, creativity and other brain functions are being used as inspirations for computational algorithms . The connectome project  aims at filling the gaps in our knowledge about the information flow in the brain. The involvement of numerous cortical and subcortical brain areas in information processing is already partially known. This should allow for creation of brain-based representations of linguistic concepts. Other goals of neurocognitive informatics include models of spreading neural activation and priming effects in psychology, especially their influence on associative thinking; the transition from conscious to unconscious control in learning new skills; the process of learning a profession, from novice to an expert; understanding and modeling of various forms of imagery and its connections to creativity and talent; development of basic concepts in infant brains, understanding of affective states, and the effects of various dysfunctions at the low-level (neural ionic channels, degeneration) on mental functions. </para>
      <para id="id1166118839989">For the first time in history the mind-body problem has solid scientific foundations based on correlations between brain states, as measured by its metabolic or electrical activity, and mental states (intentions, perceptions). The brain-mind transformation essentially requires transformation of neural activity <emphasis effect="italics">x</emphasis><emphasis effect="italics">i</emphasis>(<emphasis effect="italics">t</emphasis>), for example measured by <emphasis effect="italics">i</emphasis>=1..<emphasis effect="italics">n</emphasis> EEG electrodes, into mental activity <emphasis effect="italics">m</emphasis><emphasis effect="italics">j</emphasis>(<emphasis effect="italics">t</emphasis>), where each dimension in the “mental space” is related to some quality of our inner experience, such as intentional feelings, tactile, auditory, visual and other sensory features. </para>
    </section>
    <section id="id6331259">
      <title>Large-scale NLP projects.</title>
      <para id="id4031610">This section presents inspirations for natural language processing based on current understanding of the neuroscience of language. The goal is to connect neural processes responsible for language processing with more traditional ways of analysis, such as semantic networks and vector models, in hope that this will lead to useful enhancement of existing models. J. Feldman in his book on a “Neural Theory of Language” (NTL)  stressed two points: “Thought is structured neural activity”, and “Language is inextricable from thought and experience”. Many current attempts to create large-scale semantic networks, ontologies, and lexical resources ignore these facts, trying to learn everything from statistical correlations that may be found in the text. Lexical resources, such as WordNet<footnote id="id4638407">http://wordnet.princeton.edu</footnote>, has been described as “a lexical reference system whose design is inspired by current psycholinguistic theories of human lexical memory. English nouns, verbs, adjectives and adverbs are organized into synonym sets, each representing one underlying lexical concept. Different relations link the synonym sets” . This is certainly a step in good direction, but the definitions and properties of concepts provided by Wordnet are very brief and do not allow to understand the meaning of such concepts. For example, “a ball” in one of the senses is defined as “a spherical object used as a plaything”, and “plaything” is defined as “an artifact designed to be played with”. A ball has many properties that we know from experience: size, color, softness or hardness, surface texture, some balls may be held in hand, or kicked, or bounced on the head, or hit with a stick. In fact “a ball” is a high level concept that subsumes many experiences. A detailed representation of many specific types of balls is stored in our brain and activated by the context in which this word appears. We know that the golf ball is smaller than the cricket ball, that a small steel ball is called a shot, and that in some languages (for example, Polish) a bullet is called a ball, even if it is not quite round. Even for such trivial example any human could build a whole ontology of different types of balls. The information in Wordnet, which is the largest hand-crafted project of its kind, with more than 200,000 words-sense pairs, is minimal, not sufficient to understand any concept, ask intelligent questions to get precise information about the concept, and use it meaningfully in a discussion. </para>
      <para id="id4111869">In recent year many large-scale projects have been formulated, hoping to collect sufficient linguistic information for NLP systems. In the internet-based collaborative Open Mind Common Sense Project over 15,000 authors typed since the year 2000 over 700,000 sentences with all kind of obvious “common sense” facts. The ConceptNet knowledgebase derived from this project  has 200,000 assertions and the full base contains 1.6 million assertions. These assertions include spatial, physical, social, temporal, and psychological aspects of everyday life, capturing many commonsense concepts and relations in a large semantic network with over 150.000 nodes. The Commonsense Computing project at MediaLab (MIT) plans to use this knowledge and acquire more through collaborative projects. The Open Mind Commons<footnote id="id1166118306769">http://commons.media.mit.edu/en/</footnote> is an application designed to teach the system new information and refine existing knowledge. The project is growing and has also a databases in Portugese. The concept of “a ball” is characterized here a bit better than in Wordnet: a ball is round, circular, a sphere, is a toy, can bounce, roll, break window, is being thrown, is used in football, basketball, soccer, bowling, golf, snooker, sport, … There are many statements like “A ball can sit on top of a VCR”, or “You are likely to find a ball in the mouth of a puppy”, which are not informative, and many contradictory statements: “the globe is a ball”, “the globe is not a ball”. Altogether in March 2009 there were 171 statements that mentioned the ball. This information is much richer than what is found in the Wordnet, but still contains only a small fraction of what ordinary people know about “the ball” concept. </para>
      <para id="id1790314">Several interesting projects based on ConceptNet have been formulated. Mindpixel<footnote id="id5285906">http://en.wikipedia.org/wiki/Mindpixel</footnote> project tried to create a knowledgebase by validating true/false statements and probabilistic propositions. Multi-Lingual version of ConceptNet semantic network has been used in LifeNet, based on concept relationships. It should learn from people who play simple online games, and also learn from sensors to perform activity recognition. The system should function as the first-person spatio-temporal reasoning engine, predicting and commenting events in a typical person's life. Although it looked very atractive the project has not progressed in recent years. It has also used result of the Honda Open Mind Indoor Common Sense project, based on the same principle, based on the same principle as ConceptNet, but focus on information that indoor mobile robots should know to have commonsense and recognize objects and indoor scenes. Examples given at the project webpage<footnote id="id5488452">http://openmind.hri-us.com/login.jsp</footnote> are: “coffee is made in a coffee maker which is in a kitchen; to find out if it is raining one needs to look out of an open window; master bedroom usually has an attached bath”. </para>
      <para id="id3633655">Other important large-scale project that goes beyond WordNet ideas of a lexical database in which semantic relations, such as synonymy, antonymy, hyponymy, hypernymy and meronymy help to define meaning, is HowNet . This project<footnote id="id5890702">http://www.keenage.com</footnote> has been in development for over 20 years. Hownet stresses not only ontological, but also temporal aspects of concepts, and has its own structure of inter-concept and inter-attribute relations. The idea of a sememes, smallest basic units of meaning, is rather natural in Chinese, as the ideographical characters used for writing refer to such units. Analysis of 6000 Chinese characters led to about 1600 sememes that have been used to define about 100.000 concepts for English, and roughly the same number of Chinese language. While WordNet provides barely enough information to distinguish between concepts HowNet is more constructive, adding not only superordinate terms (hypernyms), but many more specific dependency relations, for example attribute-host, material-product, event roles, dynamic role, concept co-occurrence, etc. The Knowledge Dictionary stores the main and the secondary features of concepts, synonyms, antonyms and converse relations, event-relatedness and role-shifting information, organized in supplementary databases. Still this database is very limited, with a single attribute for a ball, shape=round. </para>
      <para id="id1166113434739">The FrameNet project<footnote id="id4418854">http://framenet.icsi.berkeley.edu/</footnote>, developed at Berkeley, creates a rich on-line lexical resource for English and Spanish, based on the frame semantics . The third release (mid-2008) of this lexical database contained more than 10,000 lexical units (word-meaning pairs), over 60% fully annotated in more than 825 semantic frames and demonstrated in over 135,000 annotated sentences. This exercise is aimed at illustrating all valid ways of using each word in each of its senses. For example, the verb bake is linked to 3 frames: Apply_heat, Cooking_creation, and Absorb_heat, depending on the word usage. This approach has its roots in Case Grammars , replacing a small number of fixed roles with large number of semantic relations empirically derived from annotation of texts. FrameNet has focused more on verbs than nouns and therefore is not suitable as an ontology of things, but FrameNet has been linked to Suggested Upper-Merged Ontology (SUMO)<footnote id="id1166118426136">http://www.ontologyportal.org </footnote>. FrameNet has developed a visualization tool for viewing the relations between frames. Connections between frames and their elements may be visualized. Automatic labeling of text with semantic frame information could greatly facilitate machine translation and semantic search, but so far are done mostly in manual way. This approach may be quite successful in specialized domains. A trilingual (English, German, French) lexical resource called The Kicktionary , describes the language of soccer. It combines frame semantics and WordNet style semantic relations, with around 2,000 lexical units organized in 104 frames and 16 scenarios, illustrated by multilingual corpus of soccer match reports. </para>
      <para id="id1166114418231">The MindNet project at Microsoft Research<footnote id="id7852552">http://research.microsoft.com/en-us/projects/mindnet/default.aspx</footnote> is aimed at building in a fully automatic way semantic networks from dictionaries, encyclopedias, and free text. Input sentences are parsed, semantic dependency graphs built, and many individual graphs aggregated into a single large graph. Probabilistic weights are assigned to subgraphs depending on their frequency in the corpus. This is an interesting way of acquiring world knowledge on the scale needed to support common-sense reasoning. Relations between concepts may be traced using an on-line browser that shows intermediate concepts, frequent collocations or similarity of paths in the subgraphs created by the system. MindNet has been used in machine translation projects at Microsoft but the last presentation has been in 2005. </para>
      <para id="id1166112527737">These project represent large-scale efforts to collect linguistic data from free texts and structured sources of information, using automatic or labor-intensive algorithms. After a period of enthusiasm many projects slow down and despite brilliant ideas seem to come to a standstill. None of these projects has tried to understand how human brains implement language, how can we know so much by learning so little? The short answer is: language describes inner activation of brain areas, that is experience. </para>
    </section>
    <section id="id5262400">
      <title>Language and memory</title>
      <para id="id1166113545844">We intend to approach NLP using inspirations from general understanding of the human information processing, investigate the feasibility of text understanding based on neurocognitive inspirations, namely representation of words in the brain and the use of various brain memory systems. Neurocognitive approach to NLP follows inspirations from brain science focusing on approximated models of memory and other neural processes. The long-term goal is to reach human-level competence in natural language processing. </para>
      <para id="id7617544">Schank was right in his early “conceptual dependency” theory of understanding language  claiming that language cannot be studied without taking memory into consideration. Analysis of texts, independent of the purpose, requires three main steps, that include at least: </para>
      <list id="id6416017" list-type="bulleted">
        <item>mapping from strings of letters to unique terms, recognition of tokens; </item>
        <item>grouping terms into phrases, mapping them to concepts and resolving ambiguities; </item>
        <item>semantic representation of the whole text, capturing relations among entities that are involved, facilitating inferences, and thus allowing for understanding and answering questions about its content. </item>
      </list>
      <para id="id3905034">These three steps roughly correspond to functions of four kinds of human memory : recognition, semantic, episodic and working memory. Semantic networks  have been explicitly inspired by the theory of semantic memory, but NLP research largely ignored these inspirations, focusing on formal approaches . </para>
      <para id="id1166114418518"><emphasis effect="bold">Recognition memory</emphasis> helps to ignore most spelling errors. Context and anticipation plays a major role in correct recognition. As long as the first and the last letter of the word is not changed even severely distorted texts containing <emphasis effect="italics">wrods wtih many paris of letres trasnpoesd</emphasis> is read without much trouble, a phenomenon that is of interest to spammers and cognitive scientists. Unstructured free texts, and especially conversational texts need a lot of data cleaning. This is traditionally done by spellcheckers, but without understanding the context may lead to wrong recommendations. For example, Lexical Systems Group of the US National Library of Medicine has developed a spelling suggestion tool called Gspell<footnote id="id3672755">http://lexsrv3.nlm.nih.gov/SPECIALIST/Projects/gSpell/V0.0.40/GSpell.html</footnote> that is based on the SPECIALIST lexicon containing general English words and many biomedical terms. Medical terminology is difficult to spell for the average English speaker (“14% of all queries submitted for health information retrieval contain a misspelled term”, ). Gspell does not use context to understand the topic and makes many spelling suggestions, although humans “see” only a single term, frequently paying no attention to the misspellings. Reading text leads to priming effects in the brain, creating expectations and anticipations for a few selected words, and inhibition of many others that “do not come to our mind”. Recognition memory cannot be separated from other memory systems, doing much more than just searching for similar terms in the lexicon. </para>
      <para id="id1927503">According to Hecht-Nielsen’s confabulation theory of the brain functions  generating expectations is almost all that neocortex does. Statistical language processing models applied to a very large text corpus used for training allowed him for prediction of the next word in a sequence with high reliability, partially capturing this anticipation, although statistical algorithms do not approximate well real brain processes behind this phenomenon. Anticipation may help to disambiguate word senses, facilitating the mapping of terms into concepts. Although his model has not been directly used for spell-checking that includes context it will lead to a quite similar approach to that taken by Google in “Google suggest”, that is based on statistics of longer n-grams. Local context is also included in another tool also called Gspell (for MacOS only), a spelling checker that searches Google to find spelling recommendations, including personal names, site names and other words that are not found in standard dictionary. A better approximation to functions of biological recognition memory should create possible candidate words that could result from various misspellings, and remove (inhibit) those that do not fit to the current context, defining the topic of discourse, or an episode. This will be elaborated further below.</para>
      <para id="id1166118642928"><emphasis effect="bold">Semantic memory</emphasis> encodes in the activity of brain's subnetworks factual information about objects and concepts, together with their properties and relations, general knowledge of the world largely independent of context and personal relevance. It is thus an interesting inspiration for knowledge representation. Formal models of semantic networks inspired by psychological ideas about semantic memory ,, have been developed in the last three decades in artificial intelligence ,. These models were created at the time when not much has been known about neurobiology of semantic memory and thus do not try to approximate functions of biological semantic memory (SM). Psychologist simply postulated nodes that correspond to concepts and different type of links that connect them, without thinking about the actual implementation of these processes in the brain. Such structure of semantic network does not take into account the fact that each node in such network is in fact a neural circuit, with similarities and associations between concepts resulting from sharing some common elements or mutual activations that are responsible for semantic priming. </para>
      <para id="id1166113625501">The <emphasis effect="italics">semantic priming</emphasis> (SP) phenomenon has been known in cognitive psychology since more than 30 years, and has been extensively studied in experimental psychology (see the review by McNamara ). Each word excites brain subnetworks that encode different meanings of that word . In such coding identical phonological representations of words may be shared among several concepts without leading to any problems. Words that have been processed earlier (context) have already activated many brain subnetworks, increasing the probability of a particular meaning of the new concept, and inhibiting all other meanings. This competition, leading to inhibition of subnetworks coding alternative meanings of the word, makes it hard to think about alternatives when one of the meanings (interpretations) fits really well to the current context. These alternatives simply “do not come to our minds”. Semantic memory develops slowly as a result to learning, forming a view of the world that reflects frequently repeated situations and regularities. Training neural networks associations between facts, description of properties of various objects, captures similarity of these objects in the activations of hidden neurons (corresponding to association cortex, connected to sensory and motor areas). So far this process has been modeled only on a toy example , but may provide a bridge towards practical NLP algorithms based on vector representations of the activations of different parts of the brain. Recent results with fMRI neuroimaging of brain states  show that when the subject thinks, hears or sees an image invoking an object similar activity of the brain arises even in different people. Prototype probability distributions of activation of different brain areas may thus serve as a brain-based vector representation that approximates this process.</para>
      <para id="id7122273">Semantic memory is frequently not sufficient for disambiguation of concept meanings because a wider context may be needed. This is provided by the <emphasis effect="bold">episodic memory</emphasis> that includes times, places, associated emotions, and other contextual knowledge of autobiographical events, incorporating also current situation, relations between objects, participants, understanding the topic of the text fragments and discourses. In short, this type of memory encodes who did what to whom where and when. Episodic level of encoding and processing information facilitates selection of interpretations at the semantic and recognition level. All three levels are in fact strongly coupled, because all brain areas are strongly coupled, creating in a fraction of a second one dynamical state. A few computational models of general episodic memory exist -, with many others focused on the role of hippocampus . In the TraceLink model  hippocampus provides links that activate cortical columns, forming coherent quasi-stable attractors corresponding to episodic states. These links may serve as episodic memory representation, an alternative to distribution of activations over columns or larger regions of the brain. </para>
      <para id="id3508076"><emphasis effect="bold">Working memory</emphasis> also plays important role in forming episodic memory. Initially it has been postulated as a theoretical psychological construct by Baddeley and Hitch , serving as a gateway between sensory input long-term memory. Neuroimaging experiments and lesions studies showed involvement of prefrontal and parietal cortex, anterior cingulate, and parts of the basal ganglia. This is a rather complicated network of diverse structures, actively involved in learning skills, providing temporal storage for manipulation of information and storing relevant information permanently . Part of the content of working memory is experienced in a conscious way and plays key role in the reinforcement learning of skills . Biologically-inspired cognitive architectures, such as Leabra  provide models of prefrontal cortex capable of short term storage that is preserved despite on-going short-term memory activity. ACT-R and other symbolic architectures treat working memory simply as a collection of elements (rules, conditions) that have actually been recalled and are available for use . </para>
      <para id="id8057260">Different knowledge representation schemes suitable for different applications may be formulated starting from such inspirations. Creation of large-scale semantic networks, and coupling with models of episodic memory for topic categorization and recognition memory for expansion of abbreviations, acronyms and misspellings is going to be of fundamental importance in all NLP-related areas. However, more detailed analysis of language-related brain activity will be even more inspiring.</para>
    </section>
    <section id="id4376328">
      <title>Words in the brain and the role of non-dominant hemisphere</title>
      <para id="id5673287">How are words and concepts represented by brains? Acoustic speech input is quickly changed into categorical, phonological representation. This representation helps to understand language in a voice-independent way, but also make understanding of languages that have quite different phonological structure from our native language rather difficult . A small set of phonemes is linked together in ordered string by a resonant state representing word form, and extended to include other brain circuits defining semantic concept. Hearing a word phonological processing creates localized attractor state whose activity spreads in about 90 ms to extended areas defining its semantics . To recognize a word in a conscious way activity of its subnetwork must win a competition for an access to the working memory -. Hearing a word activates strings of phonemes priming (decreasing the threshold for activity) all candidate words, as well as some non-word phoneme combinations. Polysemic words may use a single phonological representation that differs only by semantic extension. Unless attention is paid to the form of the word (decreasing activation thresholds for phonological representations), as when puns are deliberately made, there is little interference from other meaning of the word thanks to the inhibitory processes. This point of view agrees with the cell assembly model of language that has already quite strong experimental support and agrees with broader mechanisms responsible for memory. In the cell assembly (or neural clique) model words (or general memory patterns) are represented by strongly linked subnetworks of microcircuits that bind articulatory and acoustic representations of spoken words. The meaning of the word comes from extended network that binds related perceptions and actions, activating sensory, motor and premotor cortices. Various neuroimaging techniques confirm existence of such semantically extended networks. </para>
      <para id="id4071194">In people who can read and write visual representation of words in the recently discovered Visual Word Form Area (VWFA) in the left occipitotemporal sulcus is strictly unimodal . Adjacent Lateral Inferotemporal Multimodal Area (LIMA) reacts to both auditory and visual stimulation and has cross-modal phonemic and lexical links . It is still controversial, but the auditory word form area may also exists . It may be a homolog of the VWFA in the auditory stream, located in the left anterior superior temporal sulcus; this area shows reduced activity in developmental dyslexics. Such representation of words should help to focus symbolic thinking. Context priming selects extended subnetwork corresponding to a unique word meaning, while competition and inhibition in the winner-takes-all processes leaves only the most active candidate network. Semantic and phonological similarities between words should lead to similar patterns of brain activations for these words.</para>
      <para id="id1166112391270">Language cannot be separated from memory and thinking processes. In the dominant, usually left hemisphere (LH), phonological representations may encode precisely word forms in local circuits capable of strong activations (as differences between words and non-words in evoked response potentials  seem to indicate), and thus quickly activate extended networks that make words meaningful. The role of contralateral hemisphere, usually the right one (RH), in language processing is usually reduced to analysis of prosody and non-verbal cues, but there is growing evidence that it is also involved in language interpretation . Additional information on the role of right hemisphere comes from studies of the sudden insight (Aha!) experience that accompanies solutions of some problems. </para>
      <para id="id7525309">In experiments with functional MRI and EEG problem solving situations that have been accompanied by insight have been contrasted with those when step-by-step gradual solution did not required insight ,. About 300 ms before the moment in which Aha! experience was felt a burst of gamma activity was observed in the Right Hemisphere anterior Superior Temporal Gyrus (RH-aSTG). The delay of 300 ms is consistent with studies showing delay of subjectively felt conscious experience in respect to the brain activity . The gamma burst has been interpreted as “making connections across distantly related information during comprehension (…) that allow them to see connections that previously eluded them” . Bowden <emphasis effect="italics">et al.</emphasis>  performed a series of experiments that confirmed the EEG results using fMRI techniques. One can conjecture that this area is involved in higher-level abstractions that can facilitate indirect associations . The initial impasse in problem solving is due to the inability of the processes in the left hemisphere, focused on the precise representation of the problem, to make progress in finding a solution. Solution means that from initial configuration <emphasis effect="bold">S</emphasis><sub>0</sub> of neural activations representing the problem, a series of transitions states <emphasis effect="bold">S</emphasis><emphasis effect="italics">k</emphasis>, each with phonological component (symbolic interpretation), may be generated, until desired configuration <emphasis effect="bold">F</emphasis>=<emphasis effect="bold">S</emphasis><emphasis effect="italics">n</emphasis> is reached. The requirement for symbolic interpretation means that intuitive solutions, when we feel that some conjecture is true but are not able to justify it, may be correct but are not sufficient. </para>
      <para id="id1166118468114">The RH has only an imprecise view of the left hemisphere (LH) activity, generalizing over similar concepts and their relations. This activity represents abstract concepts, corresponding to categories higher in ontology, but also captures complex relations among concepts, relations that may not have symbolic name, but are useful in reasoning and understanding. Sometimes we feel that something is quite out of context, or that the sentence we are producing is hard to complete in grammatical way, and this may be reaction of language-related areas in the right hemisphere. For example, “left kidney” sounds correct, but “left nose” seems strange, although we do not have a concept for spatially extended things that “left” applies to. The feeling arising from understanding words and sentences may be connected to the left-right hemisphere activation interplay. Most of RH activations do not have phonological components; the activations result from diverse associations, temporal dependencies and statistical correlations that create certain expectations. Current view of recognition memory distinguishes two functionally distinct processes, recollection and familiarity (dual process theory ), giving rise to distinct phenomenal states. Familiarity-based memory leads to a feeling of knowing without reference to specific information. Conclusion of a recent paper  where familiarity with words presented 10 minutes before was tested are: “Across a variety of brain regions, the neural signature of recollection was found to be distinct from familiarity, demonstrating that recollection cannot be attributed to familiarity strength”. The effects in the left hemisphere were stronger, “but in most cases, similar but weaker activity was observed in the right hemisphere”. One can assume that interpretation of speech or text is greatly enhanced by “large receptive fields” in the RH, which can constrain possible interpretations, help in the disambiguation of concepts and provide ample stereotypes and prototypes that generate various expectations.</para>
      <para id="id4496081">Impasse in problem solving may be overcome by decreasing activations in the verbal language areas (when the conscious efforts to solve the problem are given up) and trying to solve the problem at the non-verbal, intuitive level. Activations of larger subnetworks in the RH may be sufficient to connect the starting and the desired final configurations <emphasis effect="bold">S</emphasis><sub>0</sub> and <emphasis effect="bold">F</emphasis>, creating a strong resonance state that will give rise to a gamma burst. Sometimes it leads to a brief feeling that the solution is imminent, although it has not yet been formulated in symbolic terms, a common feeling among scientist and mathematicians. The LH impasse is removed when relevant activations are projected back from the less-focused right hemisphere, allowing new dynamical associations to be formed among brain subnetworks primed by the problem description. High-activity gamma burst is projected to the left hemisphere prime LH subnetworks with sufficient strength to activate intermediate configurations, making the transition from <emphasis effect="bold">S</emphasis><sub>0</sub> through <emphasis effect="bold">S</emphasis><emphasis effect="italics">k</emphasis> to <emphasis effect="bold">F</emphasis> possible, creating associative connections linking the problem statement with partial or final solution. An emotional component is needed to increase the plasticity of the brain and remember these associations. </para>
      <para id="id1166118797161">The “Aha!” experience may thus result from the activation of the left hemisphere areas by the right hemisphere, with a gamma burst helping to bring relevant facts to the working memory, making them available for conscious processing. The final step in problem solving requires synchronization between several left hemisphere states, representing transitions from the start to the goal through intermediate states. This seems to be a universal mechanism that should operate not only in solving difficult problems, but also on much shorter time scale in understanding of complex sentences. The feeling “I understand” signifies the end of the processing and readiness of the brain to receive more information.</para>
      <para id="id1405396">Sigmund Freud is frequently quoted: “When making a decision of minor importance, I have always found it advantageous to consider all the pros and cons. In vital matters, however, the decision should come from the unconscious, from somewhere within ourselves.”. The Nijmegen Unconscious laboratory<footnote id="id7653062">http://www.unconsciouslab.com/index.php</footnote> at Radboud University focuses on issues unconscious thought, and their results agree with this statement, pointing to the important role of the right hemisphere in thinking and decision making. </para>
    </section>
    <section id="id1166114854252">
      <title>Creativity in language comprehension and production</title>
      <para id="id1166112561848">Understanding language is not only a process based on logical inferences but requires also imagination and creativity. For example, hearing new words or novel expressions that cannot be found in the dictionary we may usually guess their sense. Producing non-trivial sentences, surprising analogies or figures of speech may also be a creative act. Creativity research is done mostly by psychologist, with a few exceptions , where EEG studies have shown greater neural activity in the right parieto-temporal areas, higher alpha activity during impasse or inspiration period, and greater tendency for emotional excitement, all of which agrees with remarks on the role of right hemisphere presented in the previous section. New results obtained have been reported only very recently by the “positive neuroscience” group  that studies creativity using multimodal imaging techniques. A network of brain regions involved in creative thinking has been identified, comprised of the anterior cingulate gyrus, superior temporal gyrus, and the corpus callosum. </para>
      <para id="id7121758">Relationships between creativity and associative memory processes have been noticed a long time ago . The pairwise word association technique is perhaps the most direct way to analyze associations between subnetworks coding different concepts. These associations should differ depending on the type of priming (semantic or phonological), structure of the network coding concepts, the activity arousal due to the priming (the amount of energy pumped into the resonant system). In a series of experiments  phonological (distorted spelling) and semantic priming was applied, showing for a brief (200 ms) moment the priming cue (word) before the second word of the pair was displayed. Two groups of people, with high and low scores in creativity tests were participating in this experiment. Two type of associations were presented, simple (close) and difficult (remote), and two types of priming, positive (either phonological or semantic relation to the second word) and neutral (no relation). Creative people should have greater ability to associate words and should be more susceptible to priming. Less creative people may not be able to make remote associations at all, while creative people should show longer latency times before noticing such associations or claiming their absence. This is indeed observed, but other results have been quite puzzling . </para>
      <para id="id1166118847822">Neutral priming, based on the nonsensical or unrelated words, increased the number of claims that words are related, in case of less creative people stronger than positive priming, and in case of more creative people in a slightly lower way. Phonological priming with nonsensical sounds partially activates many words, adding intermediate active configurations that facilitate associations. If associations between close concepts are weak neutral priming may activate intermediate neural oscillators (pumping energy to the system, increasing blood supply), and that should help to establish links between paired words, while positive priming activates only the subnetwork close to the second word, but not the intermediate configurations. For creative people close associations are easy to notice and thus adding neutral or positive primes has similar small effect. Situation is quite different for remote associations. Adding neutral priming is not sufficient to facilitate connections in less creative brains when distal connection are completely absent, therefore neutral priming may only make them more confused. Adding some neural noise may increase the chance to form resonance state if weak connections exist in more creative brains – in the dynamical systems language this is called the stochastic resonance phenomenon . On the other hand adding positive priming based on spelling activates only phonological representations close to that of the second word, therefore there is no influence. Priming on positive (related) meaning leads to much wider activation, facilitating associations. These results support the idea that creativity relies on the associative memory, and in particular on the ability to link together distant concepts.</para>
      <para id="id5374880">Problems that require creative solutions are difficult to solve because neural circuits representing object features and variables that characterize the problem have only weak connections, and the probability of forming appropriate sequence of cortical activities is very small. The preparatory period – reading and learning about the problem – introduces all relevant information, activating corresponding neural circuits in the language areas of the dominant temporal lobe, and recruiting other circuits in the visual, auditory, somatosensory and motor areas used in extended representations. These brain subnetworks are now “primed”, and being highly active reinforce mutually their activity, forming many transient configurations and inhibiting at the same time other activations. Difficult problems require long incubation periods that may be followed by an impasse and despair period, when inhibitory activity lowers activity of primed circuits, allowing for recruitment of new circuits that may help to solve the problem. In the incubation period distributed sustained activity among primed circuits leads to various transient associations, most of them short-lived and immediately forgotten. Almost all of these activations do not have much sense and are transient configurations, fleeting thoughts that escape the mind without being noticed. This is usually called imagination. Interesting associations are noticed by the central executive and amplified by emotional filters that provides neurotransmitters increasing the plasticity of the circuits involved and forming new associations, pathways in the conceptual space. </para>
      <para id="id5693614">Results of experimental and theoretical research lead to the following conclusion: creativity involves neural processes that are realized in the space of neural activities reflecting relations in some domain (in case of words knowledge about morphological structures), with two essential components: 1) distributed chaotic (fluctuating) neural activity constrained by the strength of associations between subnetworks coding different words or concepts, responsible for imagination, and 2) filtering of interesting results, amplifying certain associations, discovering partial solutions that may be useful in view of the set goals. Filtering is based on priming expectations, forming associations, arousing emotions, and in case of linguistic competence on phonological and semantic density around words that are spontaneously created (density of similar active configurations representing words).</para>
      <para id="id6008319">It is thus quite likely that language comprehension and creative processes both require subsymbolic models of neural processes realized in the space of neural activities, reflecting relations in some experiential domain, and therefore cannot be modeled using semantic networks with nodes representing whole concepts, nor other techniques, that have been used in the large-scale projects described above. </para>
    </section>
    <section id="id1166112567927">
      <title>Neurocognitive informatics approach to language</title>
      <para id="id1166118609154">Neurolinguistic observations described above may be used as inspirations for neurocognitive models. The same neural processes should be involved in sentence understanding, problem solving and creative thinking. In the past there have been only a few attempts to use inspiration from brain research in NLP. Lamb defined the goal of neurocognitive linguistics as: “an attempt to understand the linguistic system of the human brain, the system that makes it possible for us to speak and write, to understand speech and writing, to think using language …” . Although his approach has been quite interesting and fruitful in understanding the neuropsychological language-related problems it had no ambition to lead to creation of algorithms for text interpretation, and it is still quite exotic in the natural language processing (NLP) community. The basic premise is rather simple: each word in analyzed text is a part of an associative network where activation spreads and the states of the networks facilitate semantic interpretation of the text, yet unraveling these “pathways of the brain” is in practice quite difficult. </para>
      <para id="id1166118512501">Connectionist approach to natural language has been introduced in the influential PDP books . Miikkulainen created DISCERN , a system for subsymbolic natural language processing, with scripts, lexicon, and memory based on Kohonen’s self-organizing networks , but despite very promising start abandoned this project. Application of constrained spreading activation techniques in information retrieval , semantic search techniques  and word sense disambiguation  have recently been made. Experience so far shows that direct attempts for large-scale neural network applications to language problems may be rather difficult and therefore a more promising direction would be to find approximations of brain activity that could be connected to the vector models, extensively used in statistical approaches to language . Below a sketch of neurolinguistic inspirations that can be used to find useful approximations to spreading of brain activity during text comprehension is given.</para>
      <para id="id1166112585702">Words activate several subnetworks, two of which will be distinguished here: one encoding the form of the word (phonological, but it may also be related to the written form in iconographic or alphabetic languages), and the second, extended over the whole brain, providing the meaning for that word. Thus approximation of the spreading activation in the brain during language processing should require at least two networks activating each other. Given the word <emphasis effect="italics">w</emphasis> = (<emphasis effect="italics">w</emphasis><emphasis effect="italics">f</emphasis>,<emphasis effect="italics">w</emphasis><emphasis effect="italics">s</emphasis>) with phonological (or visual written) component <emphasis effect="italics">w</emphasis><emphasis effect="italics">f</emphasis>, an extended semantic representation <emphasis effect="italics">w</emphasis><emphasis effect="italics">s</emphasis>, and the context <emphasis effect="italics">Cont</emphasis> (previous information that has already primed the network), the meaning of the word results from spreading activation in the left semantic network <emphasis effect="italics">LH</emphasis> coupled with the right semantic network <emphasis effect="italics">RH</emphasis>, establishing a global quasi-stationary state (<emphasis effect="italics">w</emphasis>,<emphasis effect="italics">Cont</emphasis>). This state rapidly changes with each new word received in sequence, with quasi-stationary states formed after each sentence is understood. It is quite difficult to decompose the (<emphasis effect="italics">w</emphasis>,<emphasis effect="italics">Cont</emphasis>) state into components, because the semantic representation <emphasis effect="italics">w</emphasis><emphasis effect="italics">s</emphasis> is strongly modified by the context. The state (<emphasis effect="italics">w</emphasis>,<emphasis effect="italics">Cont</emphasis>) may be regarded as a quasi-stationary wave, with its core component centered on the phonological/visual brain activations <emphasis effect="italics">w</emphasis><emphasis effect="italics">f</emphasis> and with quite variable extended representation <emphasis effect="italics">w</emphasis><emphasis effect="italics">s</emphasis>. As a result the same word in a different sentence creates quite different states of activation, and the lexicographical meaning of the word may be only an approximation of an almost continuous variation of this meaning. To relate states (<emphasis effect="italics">w</emphasis>,<emphasis effect="italics">Cont</emphasis>) to lexicographical meanings, one can clusterize all such states for a given word in different contexts and define prototypes (<emphasis effect="italics">w</emphasis><emphasis effect="italics">k</emphasis>,<emphasis effect="italics">Cont</emphasis>) for different meanings <emphasis effect="italics">w</emphasis><emphasis effect="italics">k</emphasis>. Each word in a specific meaning is thus implemented as a roughly defined extended configuration of activations of a very dense neural network, with associations between the words resulting from transition probabilities from one word to the other, that follow from the connectivity and priming of the network. This process is rather difficult to approximate using typical knowledge representation techniques, such as connectionist models, semantic networks, frames or probabilistic networks. </para>
      <para id="id5888477">The high-dimensional vector model of language popular in statistical approach to natural language processing  is a very crude approximation that does not reflect essential properties of the perception-action-naming activity of the brain . The process of understanding words (spoken or read) starts from activation of the phonological or grapheme representations that stimulate networks containing prior knowledge used for disambiguation of meanings. This continuous process may be approximated through a series of snapshots of microcircuit activations ϕ<emphasis effect="italics">i</emphasis>(<emphasis effect="italics">w,Cont</emphasis>) that may be treated as basis functions for the expansion of the state P(<emphasis effect="italics">w</emphasis>,<emphasis effect="italics">Cont</emphasis>) = Σ<emphasis effect="italics">i</emphasis> α<emphasis effect="italics">i </emphasis>ϕ<emphasis effect="italics">i</emphasis>(<emphasis effect="italics">w,Cont</emphasis>), where the summation extends over all microcircuits that show significant activity resulting from presentation of the word <emphasis effect="italics">w</emphasis>. The high-dimensional vector model used in NLP measures only the co-occurrence of words <emphasis effect="bold">V</emphasis><emphasis effect="italics">ij</emphasis> =  <emphasis effect="bold">V</emphasis>(<emphasis effect="italics">w</emphasis><emphasis effect="italics">i</emphasis>),<emphasis effect="bold">V</emphasis>(<emphasis effect="italics">w</emphasis><emphasis effect="italics">j</emphasis>) in some window, averaged over all contexts. A better approximation of the brain processes involved in understanding words should be based on the time-dependent overlap between states P(<emphasis effect="italics">w</emphasis><sub>1</sub>,<emphasis effect="italics">Cont</emphasis>) | P(<emphasis effect="italics">w</emphasis><sub>2</sub>,<emphasis effect="italics">Cont</emphasis>) = Σ<emphasis effect="italics">ij</emphasis> α<emphasis effect="italics">i </emphasis>α<emphasis effect="italics">j</emphasis>  ϕ<emphasis effect="italics">i</emphasis>(<emphasis effect="italics">w</emphasis><sub>1</sub><emphasis effect="italics">,Cont</emphasis>) | ϕ<emphasis effect="italics">j</emphasis>(<emphasis effect="italics">w</emphasis><sub>2</sub><emphasis effect="italics">,Cont</emphasis>). Systematic study of transformations between the two bases: activation of microcircuits ϕ<emphasis effect="italics">i</emphasis> and activation of complex patterns <emphasis effect="bold">V</emphasis>(<emphasis effect="italics">w</emphasis><emphasis effect="italics">i</emphasis>), has not yet been done for linguistic representations in the brain. Analysis of memory formation in mice hippocampus  in terms of combinatorial binary codes signifying activity of neuronal cliques goes in this direction. The use of wave-like representation in terms of basis functions to describe neural states makes this formalism similar to that used in quantum mechanics, although no real quantum effects are implied here.</para>
      <para id="id1166118521517">Spreading activation in semantic networks should provide enhanced representations that involve concepts not found directly in the text. Approximations of this process are of great practical and theoretical interest. The model should reflect activations of various concepts in the brain of an expert reading such texts. A few crude approximations to this process may be defined. First, semantic networks that capture many types of relations among different meanings of words and expressions may provide space on which words are projected and activation spread. Each node <emphasis effect="italics">w</emphasis> in the semantic network represents the whole state P(<emphasis effect="italics">w</emphasis>,<emphasis effect="italics">Cont</emphasis>) with various contexts clustered, leading to a collection of links to other concepts found in the same cluster that capture the particular meaning of the concept. Usually only the main differences among the meanings of the words with the same phonological representation are represented in semantic networks (meanings listed in thesauruses), but the fine granularity of the meanings resulting from different contexts may be captured in the clusterization process and can be related to the weights of connections in semantic networks. The spreading activation process should involve excitation and inhibition, and “the winner takes most” processes. Models of semantic networks used in NLP are only vaguely inspired by the associative processes in the brain and do not capture such details ,,. </para>
      <para id="id3414848">Crude approximation to the spreading activation processes leads to an enhancement of the initial text being analyzed by adding new concepts linked by episodic, semantic or hierarchical ontological relations. The winner-takes-all processes lead to inhibition of all but one concept that has the same phonological word form. Locally this may be represented as a subnetwork (graph) of consistent concepts centered around a prototype for a given word meaning (<emphasis effect="italics">w</emphasis><emphasis effect="italics">k</emphasis>,<emphasis effect="italics">Cont</emphasis>), linking it to words in the context. Such approach has been applied recently to disambiguate concepts in medical domain . The enhanced representations are very useful in document clusterization and categorization, as is illustrated using short medical texts described in the next section. Vector models may be related to semantic networks by looking at snapshots of the activation of nodes after several steps of spreading the initial activations through the network. In view of the remarks about the role of the right hemisphere, larger “receptive fields” in the linguistic domain should be defined and used to enhance text representations. This is much more difficult because many of these processes have no phonological component and thus have representations that are less constrained and have no directly identifiable meaning. Internal representations formed by neural networks are also not meaningful to us, as only the final result of information processing or decision making can be interpreted in symbolic terms (this idea can be used to model creativity at the phoneme-word level ). </para>
      <para id="id3845362">Results from fMRI studies of the brain activity associated with word meaning  cannot be used directly to create representations of concepts for practical applications. However, knowing a few experimental results and using a large corpus of words it has been possible to predict neural activity for many more words. Another approach is to list all important brain area, and try to evaluate how strongly a given word may activate them. This is now being done with emotional words, with dictionaries that label each word with several dimensions of valence. This may be very much extended, enumerating different brain areas and their functions and evaluating, in an automatic way, how strongly is a given concept associated with specific type of information processing. This should allow to create high-dimensional vector-based representations. </para>
      <para id="id5287340">Surprising efficiency of latent semantic analysis (LSA) , a simple procedure that decomposes term-document matrix into eigenvectors, has been summarized by Landauer: “The lesson I then take from LSA’s successes is that empirical association data, when sufficient to accurately induce how all of its elements are related to each other, makes learning from experience powerful enough to accomplish much, if not all, of what it does.” Context relations may partially be captured by eigenvectors, and given sufficiently large corpus various wordforms, past and present verbs, singular and plural nouns in all forms (inflections, derivations, and compounding) will be represented as closely related. However, LSA cannot be used for analysis of novel words or novel meanings generated by morphological composition, which are quite common in some languages. Antonyms are usually represented as highly similar. Also LSA does not provide universal representations of words, that we are hoping to create with brain-based approach. </para>
      <para id="id1166113696071">Other spectral techniques that are related to LSA include probabilistic latent semantic analysis, non-negative matrix factorization, and latent semantic mapping. </para>
    </section>
    <section id="id1166113783247">
      <title>Preliminary results</title>
      <para id="id1166118666209">Ideas outlined above have been used in several projects. The key problem is to find the pathways of spreading neural activation (associations) in human brain during linguistic task. The texts or information in a discourse does not provide sufficient basis for understanding. Semantic memory that captures common sense and the knowledge of an expert should be the basis of such algorithms. Relations stored in semantic memory should be dynamically weighted, depending on the priming effects by the context, thus changing the activation flow in different situations. On the formal level this may be done within the vector model or probabilistic models by changing the weights in the similarity functions, depending on the overall evaluation of the contexts, including emotional cues, goals etc. </para>
      <para id="id3922477">One simple algorithm that has been tested with very good results follows the relations that are stored in ontologies, and uses feature selection techniques to learn which relations are useful for discrimination. This is possible if many types of relations for a given concept are given. In medical domain the UMLS collection of ontologies , including the Medical Subject Headings (MeSH) , provides rich information, with many semantic types and relations between concepts. The algorithm requires a set of labeled documents – we have used summary discharges from hospital, that may be labeled by the main and the secondary disease names, but Wikipedia articles that are labeled by the subject headings may be used as training texts for other fields. The goal is to create vectors that will indicate which concepts should be added to the texts to make the meaning explicit. The algorithm for the medical texts proceeds as follows: </para>
      <para id="id8256164">1. Perform the text pre-processing steps: stemming, stop-list, spell-checking, either correcting or removing strings that are not recognized.</para>
      <para id="id5474142">2. Use MetaMap software  for concept discovery, with restrictive settings to avoid highly ambiguous results when mapping text to UMLS ontology, expand acronyms. </para>
      <para id="id3153273">3. Use UMLS relations to create first-order cosets (sets of new concepts associated with concepts found in the text that help to constraint its meaning); add only those types of relations that lead to improvement of classification results.</para>
      <para id="id2106279">4. Reduce dimensionality of the first-order coset space using feature ranking methods , but do not remove original (zero-order) features;. </para>
      <para id="id1857963">5. Repeat steps 3 and 4 iteratively to create second- and higher-order enhanced spaces. </para>
      <para id="id4704926">6. Create <emphasis effect="bold">X</emphasis>(<emphasis effect="italics">t</emphasis><emphasis effect="italics">i</emphasis>) vectors representing concepts.</para>
      <para id="id1166113793658">As a result initial texts are enriched by adding those </para>
      <para id="id2935410">Vectors <emphasis effect="bold">X</emphasis>(<emphasis effect="italics">t</emphasis><emphasis effect="italics">j</emphasis>) representing terms <emphasis effect="italics">t</emphasis><emphasis effect="italics">j</emphasis> have zero elements except for <emphasis effect="bold">X</emphasis>(<emphasis effect="italics">t</emphasis><emphasis effect="italics">k</emphasis>)=1 for <emphasis effect="italics">k</emphasis>=<emphasis effect="italics">j</emphasis> and for those terms that are in the cosets for a given term. They are highly dimensional and may be normalized to the unit length ||<emphasis effect="bold">X</emphasis>(<emphasis effect="italics">t</emphasis><emphasis effect="italics">k</emphasis>)||=1 without loss of information; any metric may then be used to compare them. They contain important terms to help distinguish documents of different categories. The non-zero coefficients of these vectors show connections between related terms. Iterative character of the algorithm leads to non-linear effects, feedback loops are strengthening some connections. In medical document categorization a single specific occurrence of a concept may be an important indicator of the document category. The Latent Semantic Analysis will miss it, finding linear combinations of terms that do not have clear semantics.</para>
      <para id="id1166113661897">This general algorithm helps to discover associations that are important from categorization point of view, and thus understanding of the text topics, or forming an episode. It helps to build semantic network with useful associations and has been used in many medical applications -. As a results short documents, such as hospital discharge summaries, in the enhanced space show strong clusterization, and concepts that appear in them are disambiguated without problems. Figure below shows how the similarity of documents, visualized using multidimensional scaling (using Ghostminer software<footnote id="id1842404">http://www. fqs.pl/ghostminer/</footnote>), looks like after the first 4 iterations of this procedure. Interesting clinical types may be identified, as some clusters show a mixture of several labels . </para>
      <figure id="id1166113740483">
        <media id="id1166113740483_media" alt="">
          <image mime-type="image/png" src="Obraz 1.png" id="id1166113740483__onlineimage" height="504" width="598"/>
        </media>
      </figure>
      <para id="id2964453">This process may also be used to follow the development of expertise in a given field. If a medical system is trained on general texts instead of real cases clusterization will not occur as these texts contain all possible combinations of concepts, and thus do not reflect probability of distributions of concepts that coherently describe observations that are encountered in a real world. Moreover, transition from a system that has learned from textbooks to a system that is an expert is slower than direct learning on properly selected cases with their clinical description. One may conclude that the current learning practices in medicine and many other fields are harmful, slowing the process of becoming an expert, therefore it takes many years after the study to gain and improve “intuitive” knowledge. Investigation of such processes using neurocognitive models of language is of great importance to education. </para>
      <para id="id2952305">A knowledge-based algorithm to create synsets and replace word concepts by sense concepts is based on the assumption that given a reference corpus with sufficient text material on a given topic the same concept senses will be expressed in almost all possible ways. In such case reading this reference corpus will activate appropriate senses linked to concepts more often then the wrong senses. The algorithm has following steps:</para>
      <list id="id6696047" list-type="enumerated" number-style="arabic">
        <item>Given a text <emphasis effect="italics">T</emphasis> create a corpus of reference texts <emphasis effect="italics">R</emphasis> on a similar topic; if there is no such corpus (textbooks, encyclopedia articles) find sufficient number of similar texts to <emphasis effect="italics">T</emphasis> and use it as the reference corpus.</item>
        <item>Initial synset list <emphasis effect="italics">L</emphasis> has count <emphasis effect="italics">N</emphasis>(<emphasis effect="italics">S</emphasis>)=0 for all synsets <emphasis effect="italics">S</emphasis> (is empty); it will be used to prepare synsets for <emphasis effect="italics">T</emphasis>.</item>
        <item>Read reference texts <emphasis effect="italics">R</emphasis>, take next word <emphasis effect="italics">u</emphasis> and check for collocations, creating concepts <emphasis effect="italics">w</emphasis>.</item>
        <item>Using Wordnet find all synsets <emphasis effect="italics">S</emphasis><emphasis effect="italics">i</emphasis>(<emphasis effect="italics">w</emphasis>), that contain <emphasis effect="italics">w</emphasis>.</item>
        <item>Add these synsets to the list <emphasis effect="italics">L</emphasis>, increasing the count <emphasis effect="italics">N</emphasis>(<emphasis effect="italics">S</emphasis>) by 1.</item>
        <item>Keep reading until the end of the reference texts, creating a list of concepts <emphasis effect="italics">w</emphasis> linked to synsets <emphasis effect="italics">S</emphasis><emphasis effect="italics">i</emphasis>(<emphasis effect="italics">w</emphasis>).</item>
        <item>Sort concepts and if more than one synset is linked to some concepts select the one with higher count number.</item>
      </list>
      <para id="id1166112565750">All concepts from the reference texts will have synsets and some new concepts may also appear, arising from reference concepts found in synsets. Reading the text <emphasis effect="italics">T</emphasis> to be annotated it is sufficient now to change all concepts to synsets. This algorithm may fail only in rare cases, when the same concept is used in the text in several meanings. Analysis of context is then necessary to distinguish these meanings. However, this happens only rarely and in many applications is not a problem at all. The quality of this approach is very high if appropriate reference texts are selected. </para>
      <para id="id1166113356716">Another line of research that is of great importance is in semantic memory acquisition. In medical domain this could be done using UMLS and huge ontologies, but for general texts comparable resources do not exist. Although large-scale projects such as the ConceptNet may collect some useful facts some common-sense knowledge is very hard to collect in this way. For example, structural properties of physical objects are quite obvious to anyone capable of seeing, and therefore are never described. One way to acquire useful knowledge is to learn how to ask questions to discriminate between objects. The 20-question game may be used as a paradigm for gaining more information in case of ambiguities that cannot be resolved automatically, as well as in case of images that should be classified, or objects that humans can describe but do not know how to characterize. It is a form of active learning. The question/answers competitions organized so far were focused on searching for answers, while equally important part, formulating intelligent questions, has been neglected. </para>
      <para id="id1166118554596">Web-based implementation of the 20-question game has been developed and used for many years<footnote id="id4057051">http://www.20q.net</footnote> and is sold in electronic toys. This implementation is based on about 500 fixed questions that have been weighted depending on their usefulness in determining answers in past games. It is quite successful, pointing to the fact that even in a rather small feature space it is possible to distinguish between concepts. This version is quite limited as in practice we are given some information in the text and cannot start asking from the beginning in hierarchical way. A proper description of concept properties may, however, be sufficient to formulate questions that bring maximum information in a given situation. For example, if there are 10 candidate words for translation a series of questions to disambiguate them may be asked, and either answered by checking the context, or answered by humans. </para>
      <para id="id1166118648661">A challenge is thus to create an algorithm that asks intelligent questions using semantic memory for concept description. Automatic ways to collect information for semantic memory, active search for plausible information that may be related to a given concept, correct it using dialogues, and finally use it to formulate questions, play the game and identify objects has been described in a series of papers -. Among many applications object identification is worth stressing: people describing natural objects such as plants and animals rarely use the same descriptors, making the search for an image or identification of an object that is on a picture quite difficult. A system that would ask the user a minimum number of questions and help to find out what plant is shown in the image, or what particular breed of dogs they see (there are over 2000 dog breeds in a specialized catalogues) would be very useful. </para>
      <para id="id3898036">At the sub-symbolic level one may also consider the process of understanding and creation of novel words. The basic assumption is that creativity in this domain requires imagination, which is based on spontaneous combination of morphological and phonological constituents of the priming words constrained by probability distributions for different combinations of these constituents in a given language (learned from a corpus data), and filtering of interesting and novel combinations, depending on their phonological as well as associative strength – words that contain morphemes that link the m to many other words are considered to be interesting. Results of an algorithm based on these principles for creation of novel words  show that about 2/3 of the words have already been invented by people and are in use as the names of companies, products or websites, but about 1/3 is equally interesting but novel. These results give hope that also creative process in the brain can be approximated using neurocognitive ideas. </para>
    </section>
    <section id="id4370796">
      <title>How to develop neurocognitive linguistics</title>
      <para id="id5421855">All that was said above shows that looking at the brain for inspirations may be quite fruitful for NLP if suitable approximations are defined. Neurocognitive approach is the only systematic way of moving closer to the human level of competence in the linguistic area. What should be done to further develop this approach?</para>
      <list id="id5009888" list-type="bulleted">
        <item>Analysis of experimental psychology results related to priming and connecting it to models of spreading activation.</item>
        <item>Analysis of neurolinguistic data, neuroimaging experiments related to word representation. </item>
        <item>Analysis of connectome project results, with evaluation of functions each area is involved in, to define the basis for brain-based representations, estimation of word valence and may other dimensions. </item>
        <item>Defining models that approximate neural activity by probability distributions and connecting them to vector models and statistical techniques of machine learning. </item>
        <item>Defining models for activation and inhibition spreading in networks connecting nodes that store vector representations of concepts; systematic approximations to neural models based on large-scale simulations, analysis of basins of attractors that develop in neural network’s semantic layers. </item>
        <item>Developing algorithms to discover pathways of activations in human associative thinking during text comprehension, that will define weights in vector representation of concepts. </item>
        <item>Developing context-dependent concept representations on a large scale, adding such representations to each word sense in Wordnet, and to the treebanks that represent common grammatical structures.</item>
        <item>Creating and analyzing networks that represent episodes, model people, or other complex entities, by connecting concepts in self-consistent and language-independent way (creating graphs of consistent concepts). </item>
        <item>Collecting real-world information for semantic memory through analysis of ontologies, dictionaries, encyclopedia, Wikipedia, active search, visual editable representations, dialogues, computer games and collaborative projects.</item>
        <item>Creating models for different types of memory – recognition, semantic and episodic – and their interactions.</item>
        <item>Testing concept and acronym disambiguation algorithms that result from such modeling. </item>
        <item>Comparison with spectral methods, all forms of latent indexing, mapping, factorization, Bayesian and causal networks. </item>
        <item>Testing algorithms for creation and recognition of neologisms and made-up words that are fairly common in language.</item>
        <item>Applications or resulting algorithms in machine translation.</item>
        <item>Other applications: word games, semantic search, object recognition, interactive interviews, chat-based user modeling, semantic links between encyclopedia articles and other documents, automatic collection of biography, social networks and medical applications. </item>
      </list>
      <para id="id3097315">A model of word reading that includes orthographical, phonological and semantic representations has been presented in . Attractor of 10 words in the semantic layer of this model are shown below for small and large values of synaptic noise, using fuzzy symbolic dynamics mapping (K. Dobosz &amp; W. Duch, in preparation). This type of analysis may help in investigation of spreading activation and associative processes. Narrow and deep attractors trap the system and prevent activation spreading, resulting in difficulty of forming associations, integrating perception from different modalities, and shifting attention from one perception to the other. In effect neural system may repeat the word but will not associate any meanings to this word, unable to integrate it with specific context. This problem may be at the foundation of autism, where hyperspecific representations of perceptual experiences are formed and strongly focused upon. The attractor landscape is determined by the strength of synaptic connections, but also by the way it is explored. Relations between properties of neurons, their ion channels, proteins that construct these channels and genes that construct the proteins, and the influence of various neurotransmitters on neural activations are quite complex, but should allow for understanding of various Autistic Spectrum Disorders and many language-related disorders. This area presents a big challenge for neurocognitive informatics.</para>
      <para id="id3986325">
        <figure id="id3080811">
          <media id="id3080811_media" alt="">
            <image mime-type="image/png" src="graphics1.png" id="id3080811__onlineimage" height="266" width="361"/>
          </media>
        </figure>
        <figure id="id4041360">
          <media id="id4041360_media" alt="">
            <image mime-type="image/png" src="graphics2.png" id="id4041360__onlineimage" height="266" width="361"/>
          </media>
        </figure>
      </para>
    </section>
    <section id="id3695711">
      <title>Psychiatric avatar scenario</title>
      <para id="id1166113533736">Initial contact with psychiatrist takes relatively long time and may lead to misunderstandings and patient’s suspicion if the doctor does not understand him/her immediately. Some patients do not mention their hallucinations and delusions treating them as less important than emotional problems they experience. An interview with an avatar that collects information, has semantic memory and models of psychiatric diseases, thus asking relevant question leading to a summary and initial description of the problem, may not only save time, but may be easier for some patients that are introverted and do not want to open themselves up to other people, including psychiatrists. </para>
      <para id="id3247488">The application will be based on semantic memory and the 20-question game algorithm. It will also collect a free text from the patient, either in typed or spoken form. For those patients that prefer to speak sound recording may provide additional source of useful information via the voice-stress analysis techniques that can discover hidden emotions. As a preliminary step in this direction we have analyzed 6 decision trees given in the DSM IV (Diagnostic and Statistical Manual of Mental Disorders) that allow for diagnosis in specialized psychiatric domains . Asking relevant questions speeds up diagnosis, but in real application the approach should be different. Instead of collecting scripts and scenarios we should rather collect knowledge from textbooks and other sources that describe psychiatric diseases in terms of symptoms, complaints, features, medications, expert’s opinions, and use these descriptions to form semantic memory. The information collected in this way allows for understanding patient’s answers, maintaining a meaningful dialog, and narrowing down possible questions. The psychiatrists receives a digest that should help him to understand the problem better and facilitate establishing good contact with the patient. </para>
      <para id="id1166113710397">Thus the psychiatric avatar will include the following elements: </para>
      <list id="id1661178" list-type="enumerated" number-style="arabic">
        <item>Talking head (preliminary experiments based on Haptek head have been done).</item>
        <item>Semantic memory containing structured knowledge </item>
        <item>Scripts for initiation of the dialogue, ending, calling help and several other situations.</item>
        <item>Control software based on the ideas taken from the 20-question game application. </item>
        <item>Summarization of results.</item>
        <item>Machine translation to make it multilingual. </item>
      </list>
      <para id="id1166113268675">A multilingual approach to collection of information may be very helpful at police stations. It frequently happens that communication with the police is difficult or impossible. </para>
      <para id="id7967852">Another version of similar algorithm will be used in an Alzheimer Clinic (Bad Aibling, Bayern, Germany) to solicit personal stories from patients. There is good evidence that talking about family matters will strengthen the personal identity of patients and therefore preserves family contacts, improving quality of life and shorting the cognitive decline period. Talking heads that should be able to keep conversation and actively solicit information from people may play an important therapeutic role.</para>
    </section>
    <section id="id5543378">
      <title>Conclusions and wider implications</title>
      <para id="id1166118649632">General inspirations for neurocognitive linguistics have been outlined, drawing on recent experiments with priming, insight and creativity. This analysis has led to a proposal of a novel role for the non-dominant brain hemisphere: generalization over similar concepts and their relations, creation of abstract categories and complex relations among concepts that have no name, but are useful in reasoning and understanding. This activity has been partially captured in a new construct called <emphasis effect="italics">coset</emphasis> defined for a linguistic concept as a set of concepts (partially overlapping with concepts that belong to other cosets) resulting from different relations of a given concept to other concepts. Such cosets, abstract concepts that lack phonological representation, have been used to enhance representation of texts. Approximations to the process of spreading brain activations based on probability waves is too difficult to use directly in text analysis, therefore “snapshots” of this activity has been identified with vector-based concept representation commonly used in statistical approach to NLP . However, vector representation of concepts should not be reduced to statistical correlations with other concepts, as commonly done in text analysis. Snapshot of dynamic activity patterns, defining connections with other concepts, should also include structural properties of concepts, properties that rarely appear as contexts in typical texts, but may appear in textbooks presenting domain knowledge. </para>
      <para id="id3300043">Relations between spreading activation networks, semantic networks and vector models have not yet been analyzed in details. The use of background knowledge in natural language processing is an important topic that may be approached from different perspectives. Without such knowledge analysis of texts, especially texts in technical or biomedical domains, is almost impossible. Neurolinguistic inspirations may be quite fruitful, leading to approximations of processes that are responsible for text understanding in human brains, but creation of useful numerical representation of various concepts is certainly a challenge. Large-scale semantic networks and spreading activation models may be constructed starting from large ontologies. For medical applications vector representation of concepts may be created by expansion of each term that is replaced by a coset using relationships provided by UMLS ontologies. To end up with a useful representation the utility of each new relation has to be checked, or a whole class of concepts based on some specific relations may be added to the coset and then pruned to remove concepts that are not useful in text interpretation. Association rules may also be helpful here . A crude version of such approach has been presented here and already using the second-order expansions gave quite good results on a very difficult problem of summary discharge categorization. This is one of the approaches to enhance UMLS ontologies by vectors that represent these concepts in numerical way and could be used in variety of tasks. As far as we know this is the first practical algorithm that allows for a large-scale recreation of pathways of spreading brain activations in the head of an expert who reads the text. </para>
      <para id="id2186594">Finding optimal enhanced feature spaces and simplest decompositions of medical records into classes using either sets of logical rules or minimum number of prototypes in the enhanced space is an interesting challenge. “Optimal” may here depend on a wider context as the meaning of a concept depends on the depth of knowledge an expert has. For example, family physician may understand some concepts in a different way than a cardiologist, but it should be possible to capture both perspectives using prototypes. A lot of knowledge that medical doctors gain through the years of practice is frequently never verbalized. Prototypes representing clusters of documents describing medical cases may be treated as a crude approximation to the activity of neural cell assemblies in the brain of a medical expert who thinks about a particular disease. This may be observed in clusterization of these documents if a proper space is defined. Clusters found in the documents containing hospital summary discharges may be interpreted in this way, although MDS mapping to two dimensions only has to introduce many distortions. It is relatively easy to collect information about rare cases that are subject to scientific investigation, but not the subtypes of the common ones. Finding such subclusters, or identifying different subtypes of disease, is an interesting goal that may potentially help to train young doctors by presenting optimal sets of cases for each specific cluster. It could also be potentially useful in more precise diagnoses. With sufficient amount of documents optimization of individual feature weights could also be attempted. </para>
      <para id="id1166112487490">Although much remains to be done before unstructured medical documents and general web documents will be fully and reliably annotated in an automatic way, a priori knowledge certainly will be very important. Creating better approximations to the activity of the brain representing concepts and making inferences during sentence comprehension is a great challenge for neural modeling. In medical domain ontologies, relations between concepts, and classification of semantic types enables useful approximations to the neurolinguistic processes, while in general domains resources of this sort are still missing. </para>
      <para id="id6641207">References</para>
      <list id="id1166118572036" list-type="enumerated" number-style="arabic" mark-prefix="[" mark-suffix="]">
        <item id="id3178865">Doidge, N. The Brain That Changes Itself: Stories of Personal Triumph from the Frontiers of Brain Science. James H. Silberman Books, Penguin (2007)</item>
        <item id="id1166112468954">Walker, S, A brief history of connectionism and its psychological implications. In Clark, A., Lutz, R., eds.: Connectionism in Context. Springer-Verlag, Berlin (1992) 123-144</item>
        <item id="id1166113652883">Russell, S.J., Norvig, P, Artificial Intelligence. A Modern Approach. Prentice-Hall, Englewood Cliffs, NJ (1995)</item>
        <item id="id4124659">Luger, G, Artificial Intelligence: Structures and Strategies for Complex Problem Solving. Addison Wesley 2008 (6th ed)</item>
        <item id="id1166118529247">Pulvermuller, F. (2003), <emphasis effect="italics">The Neuroscience of Language. On Brain Circuits of Words and Serial Order</emphasis>. Cambridge, UK: Cambridge University Press.</item>
        <item id="id1166118589226">R. Brooks, Elephants Don’t Play Chess, Robotics and Autonomous Systems 6, 3–15, 2000.</item>
        <item id="id1166114830256">R. Brooks, L.A. Stein, Building Brains for Bodies, Autonomous Robots 1, 7–25, 1994.</item>
        <item id="id5196695">Varela, F.J. Thompson, E, Rosch, E, The embodied mind: Cognitive science and human experience. MIT Press, Cambridge, MA, USA. 1991.</item>
        <item id="id4009276">Lakoff, G, Johnson, M. (1999) Philosophy In The Flesh: The Embodied Mind and Its Challenge to Western Thought. Basic Books.</item>
        <item id="id1166118611100">Lakoff G, Nunez R, Where Mathematics Comes From? How the Embodied Mind Brings Mathematics into Being. Basic Books, New York (2000).</item>
        <item id="id3077144">L. Steels and F. Kaplan. AIBO’s first words: The social learning of language and meaning. Evolution of Communication, 4(1):3–32, 2001.</item>
        <item id="id1166113292279">Steels, L. and De Beule, J. (2006) A (very) Brief Introduction to Fluid Construction Grammar. In Third International Workshop on Scalable Natural Language Understanding (2006).</item>
        <item id="id4585715">Hutchens, J.L.; Alder, M.D. (1998), Introducing MegaHAL, NeMLaP3 / CoNLL98 Workshop on Human-Computer Conversation, ACL (271): 274 Pulvermuller, F. (2003), The Neuroscience of Language. On Brain Circuits of Words and Serial Order. Cambridge, UK: Cambridge University Press.</item>
        <item id="id3949579">J.A. Feldman, From Molecule to Metaphor: A Neural Theory of Language. MIT Press 2006</item>
        <item id="id7069702">Gazzaniga, M. (ed.) The Cognitive Neurosciences. Cambridge, MIT Press (1995)</item>
        <item id="id1166118487305">Duch, W. Towards comprehensive foundations of computational intelligence. In Duch, W., Mandziuk, J., eds.: Challenges for Computational Intelligence. Vol. 63. Springer (2007) 261—316</item>
        <item id="id2243780">Fogel, L., Owens, A., Walsh, M., eds. Artificial Intelligence through Simulated Evolution. Wiley and Sons (1966)</item>
        <item id="id1166112492719">Goldberg, D. Genetic Algorithms in Optimization and Machine Learning. Addison-Wesley (1989)</item>
        <item id="id1166118631705">Bonabeau, E., Dorigo, M., Theraulaz, G. Swarm Intelligence: From Natural to Artificial Systems. Oxford University Press (1999)</item>
        <item>Kennedy, J., Eberhart, R., Shi, Y. Swarm Intelligence. Morgan Kaufmann (2001)</item>
        <item id="id7652117">de Castro, L., Timmis, J. Artificial Immune Systems: A New Computational Intelligence Approach. Springer (2002)</item>
        <item id="id2861353">Anderson, J.A., Rosenfeld, E. Neurocomputing - foundations of research. MIT Press, Cambridge, MA (1988)</item>
        <item id="id1166113743660">Anderson, J.A., Pellionisz, A., Rosenfeld, E. Neurocomputing 2. MIT Press, Cambridge, MA (1990)</item>
        <item id="id4925526">Kohonen, T. Self-organizing maps. Springer-Verlag, Heidelberg Berlin (1995)</item>
        <item id="id2199323">Powell, M.J.D. Radial basis functions for multivariable interpolation: A review. In Mason, J.C., Cox, M.G., eds.: Algorithms for Approximation of Functions and Data, Oxford, Oxford University Press (1987) 143--167</item>
        <item id="id1166118839274">Poggio, T., Girosi, F. Network for approximation and learning. Proceedings of the IEEE 78 (1990) 1481--1497</item>
        <item id="id1166113685351">Minsky, M., Papert, S. Perceptrons: An Introduction to Computational Geometry. MIT Press (1969)</item>
        <item id="id4447634">Kunstman, N., Hillermeier, C., Rabus, B., Tavan, P. An associative memory that can form hypotheses: a phase-coded neural network. Biological Cybernetics 72 (1994) 119--132</item>
        <item id="id4502498">Wang, D. On connectedness: a solution based on oscillatory correlation. Neural Computation 12 (2000) 131—139</item>
        <item id="id1166118638173">Duch, W, Setiono, R, Zurada, J. Computational intelligence methods for understanding of data. Proceedings of the IEEE 92 (2004) 771—805</item>
        <item id="id8297925">Gerstner, W., Kistler, W. Spiking Neuron Models. Single Neurons, Populations, Plasticity. Cambridge University Press (2002)</item>
        <item id="id3665238">Maass, W.C, M. Bishop, E. eds.: Pulsed Neural Networks. MIT Press, Cambridge, MA (1998)</item>
        <item id="id5277656">O'Reilly, R., Munakata, Y. Computational Explorations in Cognitive Neuroscience. MIT-Press (2000)</item>
        <item id="id1166113777255">Duch, W., Blachnik, M.: Fuzzy rule-based systems derived from similarity to prototypes. In Pal, N., Kasabov, N., Mudi, R., Pal, S., Parui, S., eds.: Lecture Notes in Computer Science. Volume 3316. (2004) 912--917</item>
        <item id="id4485439">Li, H., Chen, C., Huang, H.P. Fuzzy Neural Intelligent Systems: Mathematical Foundation and the Applications in Engineering. CRC Press (2000)</item>
        <item id="id2767387">Duch, W. Uncertainty of data, fuzzy membership functions, and multi-layer perceptrons. IEEE Transactions on Neural Networks 16 (2005) 10--23</item>
        <item id="id1166113458436">Nauck, D., Klawonn, F., Kruse, R., Klawonn, F. Foundations of Neuro-Fuzzy Systems. John Wiley &amp; Sons, New York (1997)</item>
        <item id="id4402911">Pal, S., Mitra, S. Neuro-fuzzy Pattern Recognition: Methods in Soft Computing Paradigm. J. Wiley &amp; Sons, New York (1999)</item>
        <item id="id7148959">Duch, W., Jankowski, N. Survey of neural transfer functions. Neural Computing Surveys 2 (1999) 163--213</item>
        <item id="id6111965">Duch, W., Jankowski, N. Taxonomy of neural transfer functions. In: International Joint Conference on Neural Networks. Vol 3, Como, Italy, IEEE Press (2000) 477--484</item>
        <item id="id7062474">Rumelhart, D.E. &amp; McClelland, J.L. (eds), (1986). <emphasis effect="italics">Parallel Distributed Processing: Explorations in the Microstructure of Cognition Vol. 1: Foundations, Vol. 2: Psychological and Biological Models</emphasis>. Cambridge, MA: MIT Press.</item>
        <item id="id2253805">Sowa J.F. (ed), (1991). <emphasis effect="italics">Principles of Semantic Networks: Explorations in the Representation of Knowledge</emphasis>. San Mateo, CA: Morgan Kaufmann Publishers. </item>
        <item>Jordan, M., T.J. Sejnowski, E. Graphical Models. Foundations of Neural Computation. MIT Press (2001)</item>
        <item id="id2979116">Zadeh, L., Kacprzyk, J. (eds) Computing with Words in Information/intelligent Systems: Foundations. Springer (1999)</item>
        <item id="id2946115">Zadeh, L. A new direction in ai: Toward a computational theory of perceptions. AI Magazine 22(1) (2001) 73--84</item>
        <item id="id2091486">Schölkopf, B., Smola, A. Learning with Kernels. Support Vector Machines, Regularization, Optimization, and Beyond. MIT Press, Cambridge, MA (2001)</item>
        <item id="id7380934">Hyvärinen, A., Karhunen, J., Oja, E. {Independent Component Analysis. Wiley &amp; Sons, New York, NY (2001)</item>
        <item id="id1166113760565">Fauconnier, G., Turner, M., 2002. The Way We Think: Conceptual Blending and the Mind's Hidden Complexities. Basic Books, New York.</item>
        <item id="id3658208">Pereira F.C, Creativity and Artificial Intelligence: A Conceptual Blending Approach. Mouton De Gruyter, 2007.</item>
        <item id="id1166118816030">Honavar, V., Uhr, L., eds. Artificial Intelligence and Neural Networks: Steps Toward Principled Integration. Academic Press, Boston (1994)</item>
        <item id="id1166113268159">Winston, P. Artificial Intelligence. 3rd ed. Addison-Wesley, Reading, MA (1992)</item>
        <item id="id6263108">Duch, W., Adamczak, R., Grabczewski, K. A new methodology of extraction, optimization and application of crisp and fuzzy logical rules. IEEE Transactions on Neural Networks 12 (2001) 277--306</item>
        <item id="id1883114">Jilk, D, Lebiere, C., O'Reilly, R, Anderson, J. SAL: An explicitly pluralistic cognitive architecture. Journal of Experimental and Theoretical Artificial Intelligence 20 (2008) 197-218</item>
        <item id="id1166113787850">Szymanski, J., Sarnatowicz, T., Duch, W. Towards avatars with artificial minds: Role of semantic memory. Journal of Ubiquitous Computing and Intelligence 2 (2008) 1-11</item>
        <item id="id7961281">Duch, W. Intuition, insight, imagination and creativity. IEEE Computational Intelligence Magazine 2(3) (2008) 40--52</item>
        <item id="id2167001">Lichtman, J., Livet, J., Sanes, J. A technicolour approach to the connectome. Nature Reviews Neuroscience 9 (2008) 417—422</item>
        <item id="id2901856">Feldman, J.A. From Molecule to Metaphor: A Neural Theory of Language. MIT Press (2006)</item>
        <item id="id1166113291242">Fellbaum C. (ed) WordNet. An Electronic Lexical Database. MIT Press 1998</item>
        <item id="id4504777">Havasi, C., Speer, R. &amp; Alonso, J. (2007) ConceptNet 3: a Flexible, Multilingual Semantic Network for Common Sense Knowledge. Proceedings of Recent Advances in Natural Languges Processing 2007</item>
        <item id="id3546054">Dong Z, Dong Q, Hownet And the Computation of Meaning, World Scientific 2006</item>
        <item id="id1166118483913">C.F. Baker, C.J. Fillmore and B. Cronin, The Structure of the Framenet Database, International Journal of Lexicography, Volume 16.3: 281-296, 2003.</item>
        <item id="id3531940">Fillmore, C.J. (1976): Frame semantics and the nature of language. Annals of the New York Academy of Sciences: Conference on the Origin and Development of Language and Speech, Vol.280: 20-32</item>
        <item id="id1166118808632">Schmidt T, The Kicktionary – A Multilingual Lexical Resource of Football Language. In: Boas, H.C. (ed.): Multilingual Framenets. New York: de Gruyter, 2009</item>
        <item id="id1166113622339">Roger C. Schank (1972). Conceptual Dependency: A Theory of Natural Language Understanding, Cognitive Psychology, (3)4, 532-631</item>
        <item id="id1166113466570">Anderson, J.R, Learning and Memory. J. Wiley and Sons, NY 1995.</item>
        <item id="id1166113802592">Manning, C.D, Schütze, H. (1999). <emphasis effect="italics">Foundations of Statistical Natural Language Processing</emphasis> Cambridge, MA: MIT Press.</item>
        <item id="id1166118838150">Crowell J, Zeng Q, Ngo L, Lacroix EM, A frequency-based technique to improve the spelling suggestion rank in medical queries. J. Am. Med. Inform. Assoc. 11(3):179-85, 2004.</item>
        <item id="id1166113432297">Hecht-Nielsen, R. Confabulation Theory: The Mechanism of Thought. Springer, 2007</item>
        <item id="id1166113779316">Lehmann F. (Ed), (1992). <emphasis effect="italics">Semantic Networks in Artificial Intelligence</emphasis>. Oxford, Pergamon.</item>
        <item id="id1818790">Collins A.M, Loftus E.F, A spreading-activation theory of semantic processing. Psychological Reviews 82, 407-28, 1975.</item>
        <item id="id1166114892797">McNamara T.P, Semantic Priming. Perspectives from Memory and Word Recognition, Psychology Press 2005</item>
        <item id="id3574610">McClelland, J.L, Rogers, T.T. The Parallel Distributed Processing Approach to Semantic Cognition. Nature Reviews Neuroscience, 4, 310-322, 2003 </item>
        <item id="id3944006">McClelland, J. L., McNaughton, B. L., &amp; O’Reilly, R. C. (1995) Why there are complementary learning systems in the hippocampus and neocortex: Insights from the successes and failures of connectionist models of learning and memory. Psychological Review, 102, 419-457</item>
        <item id="id2224379">Mitchell T.M, Shinkareva S.V, Carlson A, Chang K.M, Malave V.L, Mason R.A, Just M.A, Predicting Human Brain Activity Associated with the Meanings of Nouns, Science 320 (5880), 1191-1195, 2008.</item>
        <item id="id1655493">Meeter, M, Murre, J.M.J. (2005) TraceLink: A model of consolidation and amnesia. Cognitive Neuropsychology 22(5), 559-587</item>
        <item>Meeter, M., Myers, C.E. &amp; Gluck, M.A. (2005). Integrating incremental learning and episodic memory models of the hippocampal region. Psychological Review, 112, 560-585</item>
        <item id="id3898597">Gluck, M.A., Meeter, M. &amp; Myers, C.E. (2003). Computational models of the hippocampal region: Linking incremental learning and episodic memory. Trends in Cognitive Sciences, 7, 269-276.</item>
        <item id="id2099063">Shastri L, Episodic memory and cortico-hippocampal interactions. Trends in Cognitive Sciences, 6: 162-168, 2002.</item>
        <item>Meeter, M., Jehee, J.F.M., &amp; Murre, J.M.J. (2007). Neural models that convince: Model hierarchies and other strategies to bridge the gap between behavior and the brain. Philosophical Psychology, 20, 749-772</item>
        <item id="id1166118602448">Baddeley, A.D, Hitch, G. (1974). Working memory. In G.H. Bower (Ed.), The psychology of learning and motivation: Advances in research and theory (Vol. 8, pp. 47-89). New York: Academic Press.</item>
        <item id="id3232720">Ericsson, K. A, Kintsch, W. Long-term working memory. Psychological Review 102, 211–245, 1995</item>
        <item id="id6325626">Collette, F, Van der Linden, M, Poncelet, M, Working memory, long-term memory, and language processing: issues and future directions. Brain and Language 71(1):46-51, 2000</item>
        <item id="id3811728">Duch, W. (2005). Brain-inspired conscious computing architecture. <emphasis effect="italics">Journal of Mind and Behavior 26(1-2)</emphasis>, 1-22.</item>
        <item id="id4379465">Duch W, Oentaryo R.J, Pasquier M, Cognitive architectures: where do we go from here? In: Frontiers in Artificial Intelligence and Applications, Vol. 171 (Ed. by Pei Wang, Ben Goertzel, and Stan Franklin), IOS Press, pp. 122-136.</item>
        <item id="id4119141">Jusczyk P.W. The Discovery of Spoken Language. MIT Press 2000.</item>
        <item id="id3439267">Dehaene, S., Cohen, L. Sigman, M. &amp; Vinckier, F. (2005) The neural code for written words: a proposal. <emphasis effect="italics">Trends in Cognitive Science</emphasis><emphasis effect="italics">9</emphasis>, 335-341.</item>
        <item id="id1166118514934">Lin, L, Osan, R, Tsien, J.Z. (2006). Organizing principles of real-time memory encoding: neural clique assemblies and universal neural codes. <emphasis effect="italics">Trends in Neuroscience 29(1)</emphasis>, 48-57.</item>
        <item id="id4364884">Dehaene, S., &amp; Naccache, L. (2001). Towards a cognitive neuroscience of consciousness: Basic evidence and a workspace framework. <emphasis effect="italics">Cognition 79</emphasis>, 1-37.</item>
        <item id="id2081172">Gaillard, R, Naccache, L, Pinel, P, Clémenceau, S, Volle, E, Hasboun, D, Dupont, S, Baulac, M, Dehaene, S, Adam, C, &amp; Cohen, L. Direct intracranial, FMRI, and lesion evidence for the causal role of left inferotemporal cortex in reading. <emphasis effect="italics">Neuron 50</emphasis>, 191-204, 2006.</item>
        <item id="id1166114403572">Lindell A.K, In Your Right Mind: Right Hemisphere Contributions to Language Processing and Production. Neuropsychology Review 16(3), 131-148, 2006.</item>
        <item id="id5004987">Jung-Beeman, M, Bowden, E.M, Haberman, J, Frymiare, J.L, Arambel-Liu, S, Greenblatt, R, Reber, P.J, Kounios, J. (2004). Neural activity when people solve verbal problems with insight. <emphasis effect="italics">PLoS Biology 2</emphasis>, 500-510.</item>
        <item id="id3226662">Bowden, E.M., Jung-Beeman, M., Fleck, J. &amp; Kounios, J. (2005). New approaches to demystifying insight. <emphasis effect="italics">Trends in Cognitive Science 9</emphasis>, 322-328.</item>
        <item id="id4364545">Libet B, Freeman A, and Sutherland J.K.B. (Eds), The volitional brain: Towards a neuroscience of free will. Imprint Academic, 1999 .</item>
        <item>Just, M. A. &amp; Varma S. (2007) The organization of thinking: What functional brain imaging reveals about the neuroarchitecture of complex cognition. <emphasis effect="italics">Cognitive, Affective, &amp; Behavioral Neuroscience</emphasis> 7 (3), 153-191.</item>
        <item id="id1594404">Matykiewicz, P, Duch, W, Pestian, J. (2006). Nonambiguous Concept Mapping in Medical Domain, <emphasis effect="italics">Lecture Notes in Artificial Intelligence 4029</emphasis>, 941-950.</item>
        <item id="id3957349">Yonelinas AP: The nature of recollection and familiarity: A review of 30 years of research. Journal of Memory and Language 46:441-517, 2002.</item>
        <item id="id6062363">Yonelinas A.P, Otten L.J, Shaw K.N, Rugg M.D, Separating the Brain Regions Involved in Recollection and Familiarity in Recognition Memory. J. Neurosci. 25: 3002-3008, 2005</item>
        <item id="id1166114834879">Martindale, C, Hasenfus, N. EEG differences as a function of creativity, stage of the creative process, and effort to be original. Biological Psychology, 6(3), 157–167, 1978.</item>
        <item id="id1736110">Haier R.J, Jung R.E. (2008). Brain Imaging Studies of Intelligence and Creativity: What is the Picture for Education? Roeper Review, 30(3): 171-180.</item>
        <item id="id5495262">Mednick, S.A. (1962). The associative basis of the creative process. <emphasis effect="italics">Psychological Review</emphasis><emphasis effect="italics">69</emphasis>, 220–232. </item>
        <item id="id2135837">Gruszka, A., &amp; Nęcka, E. (2002). Priming and acceptance of close and remote associations by creative and less creative people. <emphasis effect="italics">Creativity Research Journal 14,</emphasis> 193-205.</item>
        <item id="id1166113738148">T. Wellens, V. Shatokhin, and A. Buchleitner, Stochastic resonance. Reports on Progress in Physics Vol. 67, pp. 45-105, 2004. </item>
        <item id="id1166118674283">Lamb, S. (1999). <emphasis effect="italics">Pathways of the Brain: The Neurocognitive Basis of Language</emphasis>. Amsterdam &amp; Philadelphia: J. Benjamins Publishing Co.</item>
        <item id="id3908501">Miikkulainen R. (1993). <emphasis effect="italics">Subsymbolic Natural Language Processing: An Integrated Model of Scripts, Lexicon, and Memory</emphasis>, Cambridge, MA: MIT Press. </item>
        <item id="id3891174">Miikkulainen, R. (2002). <emphasis effect="italics">Text and Discourse Understanding: The DISCERN System</emphasis>, In Dale R., Moisl H. and Somers H. (eds), A Handbook of Natural Language Processing: Techniques and Applications for the Processing of Language as Text, 905-919. New York: Marcel Dekker. </item>
        <item>Crestani, F. (1997). Application of Spreading Activation Techniques in Information Retrieval. <emphasis effect="italics">Artificial Intelligence Review</emphasis><emphasis effect="italics">11</emphasis>, 453-482.</item>
        <item id="id1166113788513">Crestani, F., &amp; Lee, P.L. (2000). Searching the web by constrained spreading activation. <emphasis effect="italics">Information Processing &amp; Management</emphasis><emphasis effect="italics">36</emphasis>, 585-605.</item>
        <item id="id7914250">Tsatsaronis, G. Vazirgiannis, M., &amp; Androutsopoulos, I. (2007) Word Sense Disambiguation with Spreading Activation Networks Generated from Thesauri, in <emphasis effect="italics">20th Int. Joint Conf. in Artificial Intelligence</emphasis> (IJCAI 2007), Hyderabad, India, pp. 1725-1730.</item>
        <item id="id1166113814436">Duch W, &amp; Pilichowski, M. (2007). Experiments with computational creativity. <emphasis effect="italics">Neural Information Processing - Letters and Reviews 11</emphasis>, 123-133.</item>
        <item id="id1166113288312">Duch, W. (2007) Creativity and the Brain. In: A Handbook of Creativity for Teachers. Ed. Ai-Girl Tan, Singapore: World Scientific Publishing, pp. 507-530. </item>
        <item id="id7600298">Itert, L. Duch, W. &amp; Pestian, J. (2007). Influence of <emphasis effect="italics">a priori</emphasis> Knowledge on Medical Document Categorization, <emphasis effect="italics">IEEE Symposium on Computational Intelligence in Data Mining</emphasis>, IEEE Press, pp. 163-170.</item>
        <item id="id1166118528848">Landauer, T. K., &amp; Dumais, S. T. (1997). A solution to Plato's problem: The Latent Semantic Analysis theory of the acquisition, induction, and representation of knowledge. Psychological Review, 104, 211-240.</item>
        <item id="id5622414">Landauer, T. K. On the computational basis of learning and cognition: Arguments from LSA. In N. Ross (Ed.), The psychology of learning and motivation, 41, 43-84, 2002. </item>
        <item id="id8293293">UMLS Knowledge Sources, 13th Edition – January Release. Available: <link url="http://www.nlm.nih.gov/research/umls">http://www.nlm.nih.gov/research/umls</link></item>
        <item id="id1850832">Medical subject headings, MeSH, National Library of Medicine, URL: <link url="http://www.nlm.nih.gov/mesh/">http://www.nlm.nih.gov/mesh/</link>.</item>
        <item id="id1166113353088">MetaMap, available at <link url="http://mmtx.nlm.nih.gov/">http://mmtx.nlm.nih.gov</link></item>
        <item id="id1838187">Duch W, Filter Methods. In: Feature extraction, foundations and applications. Eds: I. Guyon, S. Gunn, M. Nikravesh, L. Zadeh, Studies in Fuzziness and Soft Computing, Physica-Verlag, Springer, 2006, pp. 89-118</item>
        <item id="id1166113480109">Duch W, Matykiewicz P, and Pestian J, Neurolinguistic Approach to Natural Language Processing with Applications to Medical Text Analysis. Neural Networks 21(10), 1500-1510, 2008</item>
        <item>Matykiewicz P, Duch W, Zender P.M, Crutcher K.A, Pestian J.P, Neurocognitive approach to clustering of PubMed query results. In: Neural Information Proceesing, 15th Int. conference ICONIP 2008, Auckland, New Zealand, pp. 160-161, 2008.</item>
        <item>Matykiewicz P, Duch W, Pestian J.P, Clustering semantic spaces of suicide notes and newsgroups posts. ACL Conference 2009 (submitted).</item>
        <item>Duch W, Matykiewicz P, Pestian J, Neurolinguistic Approach to Vector Representation of Medical Concepts. Presented at the 20th Int. Joint Conference on Neural Networks (IJCNN), Orlando, IEEE Press, August 12-17, 2007, pp. 3110-3115</item>
        <item>Duch W, Matykiewicz P, Pestian J, Towards Understanding of Natural Language: Neurocognitive Inspirations. Springer Lecture Notes in Computer Science, Vol. 4668, 953–962, 2007.</item>
        <item>Matykiewicz P, Duch W, Pestian J, Nonambiguous Concept Mapping in Medical Domain, Lecture Notes in Artificial Intelligence, Vol. 4029, 941-950, 2006</item>
        <item id="id1166118783473">Matykiewicz P, Pestian J, Duch W, and Johnson N, Unambiguous Concept Mapping in Radiology Reports: Graphs of Consistent Concepts, AMIA Annu Symp Proc. 2006; 2006: 1024.</item>
        <item id="id1166113439850">Duch W, Szymański J, Semantic Web: Asking the Right Questions. Series of Information and Management Sciences, M. Gen, X. Zhao and J. Gao, Eds, California Polytechnic State University, CA, USA, pp. 456-463, 2008. </item>
        <item>Szymanski J, Duch W, Knowledge representation and acquisition for large-scale semantic memory. Presented at the World Congress on Computational Intelligence (WCCI'08), Hong Kong, 1-6 June 2008, IEEE Press, pp. 3117-3124</item>
        <item id="id1166118575024">Szymanski J, Sarnatowicz T, Duch W, Towards Avatars with Artificial Minds: Role of Semantic Memory. Journal of Ubiquitous Computing and Intelligence, American Scientific Publishers, 2, 1-11, 2008.</item>
        <item>Szymanski J, Duch W, Semantic Memory Knowledge Acquisition Through Active Dialogues. Presented at the 20th Int. Joint Conference on Neural Networks (IJCNN), Orlando, IEEE Press, August 2007, pp. 536-541</item>
        <item>Szymanski J, Duch W, Semantic Memory Architecture for Knowledge Acquisition and Management. Presented at the Sixth International Conference on Information and Management Sciences (IMS2007), July 1-6, 2007, California Polytechnic State University, CA, pp. 342-348</item>
        <item>Szymanski J, Sarnatowicz T, Duch W, Semantic memory for avatars in cyberspace. 2005 International Conference on Cyberworlds, Singapore 23-25 Nov. 2005, T.L. Kunii, S.H. Soon and A. Sourin (eds), IEEE Computer Society, pp. 165-171</item>
        <item id="id1166113504860">Duch W, Szymanski J, Sarnatowicz T, Concept description vectors and the 20 question game. Intelligent Information Processing and Web Mining, Advances in Soft Computing, Springer Verlag, ISBN 3-540-25056-5 (Eds. Klopotek, M.A., Wierzchon, S.T., Trojanowski, K.), pp. 41-50, 2005.</item>
        <item id="id1655154">Antonie, M.-L. &amp; Zaiane, O.R. (2002). Text document categorization by term association. <emphasis effect="italics">Proc. of IEEE Int. Conf on Data Mining (ICDM),</emphasis> pp. 19- 26.</item>
        <item>D. Roy. Semiotic Schemas: A Framework for Grounding Language in Action and Perception. Artificial Intelligence, 167(1-2):170-205, 2005</item>
        <item>D. Roy, Grounding words in perception and action: Insights from computational models, Trends in Cognitive Science, 9(8), 389-96, (2005)</item>
        <item>D. Roy, K.Y. Hsiao and Nikolaos Mavridis. (2004). Mental Imagery for a Conversational Robot. IEEE Transactions on Systems, Man, and Cybernetics, Part B, Volume 34 , Issue 3, pages 1374-1383</item>
        <item>O’Reilly, R.C. (2006). Biologically Based Computational Models of High-Level Cognition. Science, 314, 91-94</item>
        <item>J. Ruppenhofer, M. Ellsworth, M.R L. Petruck, C.R. Johnson, J. Scheffczyk, FrameNet II: Extended Theory and Practice (on-line book, 2006), URL: http://framenet.icsi.berkeley.edu/book/book.html </item>
        <item id="id3623567">Pestian, J. Aronow, B. &amp; Davis, K. (2002). Design and Data Collection in the Discovery System. In <emphasis effect="italics">Int. Conf. on Mathematics and Engineering Techniques in Medicine and Biological Science</emphasis>.</item>
        <item id="id3584625">Zhou, X., Zhang, X. &amp; Hu X., (2007). Semantic Smoothing of Document Models for Agglomerative Clustering, <emphasis effect="italics">20</emphasis><emphasis effect="italics">th</emphasis><emphasis effect="italics"> Int. Joint Conf. on Artificial Intelligence (IJCAI 2007)</emphasis>, Hyderabad, India, pp. 2922-2927.</item>
        <item id="id3901750">Cimiano, P. (2006). <emphasis effect="italics">Ontology Learning and Population from Text. Algorithms, Evaluation and Applications</emphasis>. Springer.</item>
        <item id="id7982214">Bein, W.W., Coombs, J.S., &amp; Taghva, K. (2003). A Method for Calculating Term Similarity on Large Document Collections. <emphasis effect="italics">Int. Conf. on Information Technology: Computers and Communications</emphasis>, pp. 199-207. </item>
        <item id="id1166114454206">Li, Y., Zuhair, A.B. &amp; McLean, D. (2003). An Approach for Measuring Semantic Similarity between Words Using Multiple Information Sources. <emphasis effect="italics">IEEE Transactions on Knowledge and Data Engineering</emphasis><emphasis effect="italics">15(4)</emphasis>, 871-882.</item>
        <item id="id7463104">Ghostminer data mining software, <link url="http://www.fqspl.com.pl/ghostminer/">www.fqspl.com.pl/ghostminer/</link></item>
        <item id="id7007519">Pękalska, E. &amp; Duin, R.P.W. (2005). <emphasis effect="italics">The dissimilarity representation for pattern recognition: foundations and applications</emphasis>. New Jersey; London: World Scientific. </item>
        <item id="id1166114860826">Ward, J.H. (1963). Hierachical grouping to optimize an objective function. <emphasis effect="italics">J. American Statistical Association 58</emphasis>, 236-244.</item>
        <item id="id2902942">Rousseeuw, P.J. (1987). Silhouettes: A graphical aid to the interpretation and validation of cluster analysis. <emphasis effect="italics">J. Comput. Appl. Math. 20</emphasis>, 53–65.</item>
      </list>
    </section>
  </content>
</document>
