=Paper=
{{Paper
|id=Vol-230/paper-2
|storemode=property
|title=A neural architecture for reasoning, decision-making, and episiodic memory: Taking a cue from the brain
|pdfUrl=https://ceur-ws.org/Vol-230/02-shastri.pdf
|volume=Vol-230
|dblpUrl=https://dblp.org/rec/conf/ijcai/Shastri07
}}
==A neural architecture for reasoning, decision-making, and episiodic memory: Taking a cue from the brain==
Invited keynote talk:
A neural architecture for reasoning, decision-making, and
episodic memory: Taking a cue from the brain
Lokendra Shastri
International Computer Science Institute
Berkeley, CA 94704, USA
www.icsi.berkeley.edu/~shastri
ABSTRACT
The human brain is capable of encoding a large and diverse body of common sense
knowledge and performing a wide range of inferences rapidly; to wit our ability to
understand language in real-time. Consider the simple narrative: “John fell in the
hallway. Tom had cleaned it. He got hurt.” Upon reading this narrative most of us would
effortlessly infer that Tom had cleaned the hallway, therefore, the hallway floor was wet;
John slipped and fell because the floor was wet; John got hurt because of the fall. Such
“bridging” inferences help in establishing referential and causal coherence and are
essential for language understanding. Since we understand language at the rate of several
hundred words per minute, it follows that we draw such inferences within hundreds of
milliseconds.
The ability to reason effectively with a large body of knowledge is critical not only for
language understanding, but also for decision-making, problem solving, and planning.
Consequently, the development of efficient, large-scale inference systems has been a
central goal of artificial intelligence. Although notable progress has been made toward
this goal, it remains far from being achieved.
Given that the human brain is the only extant system capable of supporting a broad range
of efficient, large-scale reasoning, it seems appropriate to assume that understanding how
the brain represents knowledge and performs inferences might lead to critical insights
into the structure and design of large-scale inference systems.
In this talk I will review the current state of a long-term research project on
understanding the neural basis of knowledge representation, reasoning, and memory, with
an emphasis on the representation and processing of relational (first-order) knowledge. In
particular, I will describe SHRUTI, a neurally plausible model that demonstrates how a
suitably structured network of simple nodes and links can encode several hundred
thousand episodic and semantic facts, causal rules, entities, types, and utilities and yet
perform a wide range of explanatory and predictive inferences within a few hundred
milliseconds. I will examine some of the predictions stemming from this work about the
characteristics of common sense reasoning, and I will discuss how insights arising from
this work can be leveraged to design a scalable inference system running on conventional
computers.