top of page

Spatial, temporal, causal and-or graph: the next step for Cognitive Science?

Searching for a unifying framework for addressing cognitive science

Cognitive science has shared parallels with artificial intelligence in the sense that it has also gone through “booming” periods, meaning times where a proposed theoretical framework seemed to be able to explain aspects of the human mind, and subsequent “busting” periods, times where that framework was rebuked by new findings[1]. Such booming and busting fluctuation revolved around the dichotomy between nativism and empiricism, with their respective ramifications[2], but regardless of this rivalry, the 21st century lifted its curtains being able to find reconciliation between these two and propose yet another theoretical framework seeking to ultimately explain how our minds work, called predictive processing (see Clark, 2013). The following text seek to defend how this framework can indeed unify several theories in the field of cognitive science and help gear future actions regarding implementing it in the field of AI.

It’s important the seeking of a unifying basis to direct the study of cognitive science because it helps make its study coherent. Furthermore, in other fields such as that of Physics, as pointed out by Prof. Song, it has been noticed that current studies of this discipline revolves around the fundamentals of forces, fields and particles. In that regard, predictive processing, along with spatial, temporal, causal and-or graphs (STC-AoG)[3] might serve as the forces, fields and particles of cognitive science.

Both predictive processing and STC-AoG converge in that they propose the existence of a mental structure that constantly processes incoming inputs, in other words, top-down deduction addressing bottom-up incoming precepts from the environment. One implication of such structure is that our minds gears our body to be able to perform complex actions and articulate complex language even with small input data, given that as little as there is, it undergoes constant processing from which predictions (stemmed from logical causal relationships established between data) are drawn. Another implication is that the mind discriminates the ubiquitous data in our surroundings, meaning, that we don’t perceive thoroughly the environment, but rather hierarchically choose data that’s beneficial for us to accomplish an immediate task.

If the above comparison of similarities hold true, then predictive processing can indeed act as a framework addressing STC-AoG, therefore help explain from the simplest of our daily life activities, while also serve as a framework for studying computational cognitive science.

For example, assuming our mind does have a representation of our immediate surrounding that follows a hierarchy and establishes causal relationships in order to make predictions and gear actions, we can then explain the action of drinking water when thirsty. For instance, check Figure 1.1[4]:



Figure 1.1: Example of STC-PG (a slight variation of the STC-AoG, see Xiong et all, 2016) in which a hierarchical-causal structure is applied[5] to the immediate environment, with the top-right-grey-pentagon indicating the purpose of being in the room known as kitchen. This helps define a series of current actions that can be taken to solve the problem of thirst, such as fill a vase with water, while also predicting later interactions with the environment depending on what is perceived, which influences future intents, such as resting, and gears future actions. Note: the camera in the lower left corner indicates that if this was a robot’s eyes, then it should be able to reconstruct this human mental structure. Image extracted from (Zhu, 2017)

On the origin of such mental representation and how it affects the design of AI

Until now, PP and STC-AoG have been discussed, but why is it that humans were born with such mental structures? A possible answer is given by Prof. Zhu, and is also mentioned by cognitive neuroscientist Anil Seth[6], which regards survivability. Perhaps it’s an innate property that humans were born with, but the pursuit of survival can provide a explanation to a lot of fundamental aspects of the mind, such as the origin of PP or the STC-AoG, that is, we have that internal predictive structure so that we can act accordingly to survive in an environment by reducing harm to our body, minimize fatigue, feed ourselves properly, maximize rewards, among others. Survivability also accounts for the origin of communication, given that cooperating with other beings and environment through language was meant to enhance survival.

Another example, now taking survivability into account, would be regarding an individual perceiving a sandstorm incoming; in order to reduce harm to his eyes, his mental structure would parse the elements in his surrounding (clothing, eyeglasses, accessories), predict their functionality and respond to the changing environment by acting (using scarf to cover eyes). Notice that he doesn’t necessarily parse everything in his surrounding, meaning that he might ignore the exact temperature of the environment, sunlight intensity, air speed, whether there’s a tree five miles away from him, among others, hence it’s a discrimination of inputs, but despite hierarchically choosing certain inputs he can still design a series of steps to protect himself. Conversely, we can ask why isn’t his mental structure guiding him to receive and withstand pain. Also notice that in this case, PP is not a registry of data as it would mean that if the individual hasn’t been exposed to the environment, then he wouldn’t know how to act as there is no previous information regarding how to behave during a sandstorm.

On Vision and Natural Language processing:

Also, PP and STC-AoG can take into account how human beings can perform complex tasks despite having a relative poor amount of input (this is partly due to the natural biologically limits of the eye[7], given that what acutely detects colour and details is in the fovea centralis, which accounts for about 15° out of 200° of range of vision (Nave, 2017)). This is by implying that humans, despite not seeing every aspect of nature in detail, instead hierarchically chooses some of them, he can predict the rest of useful features and how they can be used for their benefits. This reduces time of processing, as there is less data, and also makes actions faster. If a human were to grab a can sitting on a table, he wouldn’t place much attention in the walls of the room, as they can be internally predicted, nor even the entire table, but part of it, instead focus his attention on the can, hence he can quickly perform the task. This contrasts with how Shakey[8] behaved, for instance, given that it had to analyse each and every single piece of input independently (walls, floor, objects), without discriminating data, which then made the task of pushing cubes very slow. In that regard, both PP and STC-AoG focus more on drawing logical connections among data input, so that predictions are made, instead of quantity of data input.[9]

PP and STC-AoG can permeate into the fields of linguistics, mainly because incoming data wouldn’t be treated as unrelated individual chunks of meaningless data, but rather part of a structure that’s constantly discriminating them and drawing connections between them. For instance, regarding the discrimination of data, one can find that sentences can be set into hierarchies as show in Figure 1.3:


Figure 1.3: The way how syntax is structured perhaps displays innate principles of human language

From Figure 1.3, it’s worth noting that sentences can be broken down into parts that are more important, called the heads of a phrase, and others that are less important, the heads’ complements, hence occupy the bottom of the hierarchy, such as:



In other words, in a verb phrase, we value the verb over everything else, in a noun phrase, we value the noun, so on and so forth. Therefore, in a complete sentence, we would value nouns, verbs and adjectives, but ignore what’s considered complements to those heads. Perhaps that’s the fundamental principle of language, to describe reality by identifying an object (to which we subjectively called a “noun”), that object’s properties (to which we subjectively called an “adjective”), and any object’s function (to which we subjectively called a “verb”). This kind of discriminative parsing of utterances means less amount of data needed to be processed while also maintaining coherent speech:

George: Hi! My name is George

(What Anna’s mind processes: … name…George)

Anna: Hello, greetings from the library located next to Sainsbury, the best shop ever where you can buy anything you want, at the lowest of the lowest price.

(What George’s mind processes: …library…next Sainsbury…buy… low price)

George: what was my name again?

Anna: George, by the way, where do I work?

George: At the library, next to Sainsbury shop.

This contrasts with some of the current language processors given that they analyse data independently, without discriminating nor putting it into hierarchies, which disables coherent speech, as shown in Figure 1.4:



Figure 1.4: Screenshot of a dialogue with Cleverbot[10] displaying how the lack of a hierarchical, interconnected data input leads to incoherence in output. Each line of data is treated as a “separate” entity, instead of finding logical connections. Each word of each sentence is also given equal amount of importance.

PP also allow us to draw causal relations in sentences and build predictions out of data input. Hence, we can predict what other people might say, and even what they think based on what we have heard about them before or on data previously categorized, see Saxe & Kanwisher, 2003[11]. Consider one of the ways to test this as shown in Figure 1.5:





Figure 1.5: Participants are able to predict what Emily would think it’s a Porsche based on previous data input (Original source: (Saxe & Kanwisher, 2003))

Nonetheless, given the above example of CleverBot, it would also probably fall short when trying to predict what Emily is thinking.

Therefore, for robots to take the next step of being more human (or the next step for cognitive science), this unifying framework in the form of theory (PP) and practice (STC-AoG) can be implemented.

Alternative to natural language processing

Human language is perhaps designed arbitrarily, therefore it might be too much to program AI so that it discriminates language.

An alternative to that, provided by Prof. Zhu, is to use pictographs, as in Figure 1.7:



Figure 1.7: Pictographs can be used for communication between AI, that is, for AI to learn between one another they can use this system, while when communicating with human they use our language.


Bibliography

  1. Clark, A. (1997). Being There: Putting Brain, Body and World Together Again. Cambridge, Massachusetts: MIT Press.

  2. Clark, A. (2013). Whatever next? Predictive brains,situated agents, and the future of cognitive science. BEHAVIORAL AND BRAIN SCIENCES, 1-7, 20-21.

  3. Nave, R. (2017). The Fovea Centralis. Retrieved from HyperPhysics: http://hyperphysics.phy-astr.gsu.edu/hbase/vision/retina.html

  4. Saxe, R., & Kanwisher, N. (2003). People thinking about thinking people The role of the temporo-parietal junction in “theory of mind”. NeuroImage 19, 1835-1836, 1841-1842.

  5. Seth, A. (2017, July). Your brain hallucinates your conscious reality. Retrieved from TED Ideas worth sharing: https://www.ted.com/talks/anil_seth_your_brain_hallucinates_your_conscious_reality

  6. Xiong, C., Shukla, N., Xiong, W., & Zhu, S.-C. (2016). Robot Learning with a Spatial, Temporal, and Causal And-Or Graph. IEEE Robotics and Automation Letters (RA-L), 1-4, 8.

  7. Zhu, S.-C. (2017, 11 02). A light discussion of Artificial Intelligence: Current Situation, Purpose, Framework, and Unification | Finding the root of the issue (浅谈人工智能:现状、任务、构架与统一正本清源). Retrieved from Untitled: http://www.stat.ucla.edu/~sczhu/Blog_articles/%E6%B5%85%E8%B0%88%E4%BA%BA%E5%B7%A5%E6%99%BA%E8%83%BD.pdf


[1] This is a parallel drawn from UCLA Prof. Song Zhu-Chun’s description of the evolution of Artificial Intelligence research. [2] For empiricism, behaviourism was an example of a discipline stemmed from it. As for nativism, cognitive science emerged. [3] To see further work of Prof. Zhu, please visit VCLA@UCLA [4] The above computational implementation of the structure was done by UCLA Prof. Zhu and his team, and even though predictive processing wasn’t explicitly mentioned as their theoretical framework, it still comes with a lot of similarities. [5] Synonyms that can be used: adjusted to, adapted on. [6] He is also a proponent of predictive coding, see Seth, 2017 [7] This can be evidenced by asking individuals to draw and test whether they can exactly reproduce a landscape. Robots, through cameras, can reproduce exactly the environment, nonetheless, it would ignore the efficiency of inputting what’s necessary while predicting the rest. [8] Robot developed at the SRI International. Performance can be watched at: https://www.youtube.com/watch?v=GmU7SimFkpU [9] Prof. Song refers to “small data for big tasks” as the basis for the future of AI. [10] Software designed by British scientist Rollo Carpenter in 1986 using AI algorithm to have conversations with humans [11] Briefly, a region in the human temporo-parietal junction (TPJ-M) is involved specifically in reasoning about the contents of another person’s mind

19 views0 comments

Comentários


bottom of page