public:t_720_atai:atai-18:lecture_notes_knowledge_representations
This is an old revision of the document!
−Table of Contents
T-720-ATAI-2018 Main
Links to Lecture Notes
T-720-ATAI-2018
Lecture Notes, W9: Knowledge Representation & Meaning
Important Concepts for AGI
Autonomy | The ability to do tasks without interference / help from others / outside. | |
Reasoning | The application of logical rules to knowledge. | |
Attention | The management of processing, memory, and sensory resources. | |
Meta-Cognition | The ability of a system to reason about itself. | |
Understanding | The phenomenon of “understanding” has been neglected in AI and AGI; Modern AI systems do not understand. Yet the concept seems crucial when talking about human intelligence; the concept holds explanatory power - we do not assign responsibilities for a task who has a significant lack of understanding of the task. Moreover, the level of understanding can be evaluated. Understanding of a particular phenomenon <m>phi</m> is the potential to perform actions and answer questions with respect to <m>phi</m>. Example: Is an automobile heavier or lighter than a human? |
|
Explanation | When performed by an agent, the ability to transform knowledge about X from a formulation primarily (or only) good for execution with respect to X to a formulation good for being communicated (typically involving some form of linearization, incremental introduction of concepts and issues, in light of an intended receiving agent with a particular a-priori knowledge). Is it possible to explain something that you don't understand? |
|
Learning | Acquisition of information in a form that enables more successful completion of tasks. We call information in such a form “knowledge” or “practical knowledge”. (There is also the concept of “impractical knowledge”, which sometimes people feel must be the case of “useless trivia” that seems to be useless for anything, but can in fact turn out to be useful at any point, as for instance using such trivia to wow others with one's knowledge of trivia.) | |
Life-long Learning | Incremental acquisition of knowledge throughout a (non-trivially long) lifetime. | |
Transfer Learning | The ability to transfer what has been learned in one task to another. | |
Imagination | The ability to evaluate potential contingencies. Also used to describe the ability to predict. Relies on reasoning and understanding. |
|
Creativity | A measure for the uniqueness of solutions to problems produced by an agent; or the ability of an agent to produce solution(s) where other agents could not. Also used as a synonym of intelligence. An emergent property of an intelligent agent that relies on several of the above features. |
Representation
What it is | A representation is what you have when you pick something to stand for something else, like the lines forming the word “cup” used in particular contexts are used to represent (implicate, point to) an object with some features and properties. All knowledge used for intelligent action must have a representation. |
What it Involves | A particular process (computation, thought) is given a particular pattern (e.g. the text “cup” or the word “cup” uttered – or simply by the form of the light falling on a retina, at a particular time in a particular context) that acts as a “pointer” to an internal representation, an information structure that is rich enough to answer questions about a particular phenomenon that the “pointer” points to, without having to perform any other action than to manipulate that information structure in particular ways. |
Why it is Important | Mathematically: With the amount of information in the physical world vastly outnumbering the ability of any system to store it all in a lookup table, methods for information storage and retrieval with greater compression are needed. Historically: - The founding fathers of AI spoke frequently of representations in the first two decades of AI research. - Skinnerian psychology and Brooksian AI – both “representation-free” methodologies – largely outlawed the concept of representation from AI in the 80s onward. - Post 2000s the rise of ANNs has helped continue the trend. |
Good Regulator Theorem | Meanwhile, Conant & Ashby's Good Regulator Theorem proved (yes, proved) that every good controller (“regulator”) of a system must be a model of that system. Good Regulator Paper |
Why That Matters | A model is by definition a representation (of the thing that it is a model of). |
Bottom Line | Referring to the first table above, AGI is unlikely without sophisticated methods for representation of complex things, and sophisticated methods for their creation, manipulation, and management. This is the role of a cognitive architecture. |
Symbols & Meaning
What are Symbols? | Peirce's Theory of Semiotics (signs) proposes 3 parts to a sign: a sign/symbol, an object, and an interpretant. Example of symbol: an arbitrary pattern, e.g. a written word. Example of object: an automobile. Example of interpretant: what you see in your mind's eye when you read the word “automobile”. The last part is the most complex thing, because obviously what you see and I see when we read the word “automobile” is probably not exactly the same. |
“Symbol” | Peirce used various terms for this, including “sign”, “representamen”, “representation”, and “ground”. Others have suggested “sign-vehicle”. What is mean in all cases is a pattern that can be used to stand for something else, and thus requires an interpretation to be used as such. |
Peirce's Innovation | Detaching the symbol/sign from the object signified, and introducing the interpretation process as a key entity. This makes it possible to explain why people misunderstand each other, and how symbols and meaning can grow and change in a culture. |
Meaning | Philosophers are still grappling with the topic of “meaning”, and it is far from settled. It is highly relevant to AI, especially AGI - an AGI that cannot extract the meaning of a joke, threat, promise, or explanation is hardly worth its label. |
Current Approach | Meaning stems from two main sources. Firstly, acquired and tested models form a graph of relations; the comprehensiveness of this graph determines the level of understanding that the models can support with respect to a particular phenomenon. Meaning is not possible without (some level of) understanding. Secondly, meaning comes from the context of the usage of symbols, where the context is provided by (a) who/what uses the symbols, (b) in what particular task-environment, using ( c) what particular syntactic constraints. |
Prerequisites for using symbols | A prerequisite for communication is a shared interpretation method, shared interpretation of syntax (context), and shared knowledge (object). |
Where the Symbols “are” | When we use the term “symbol” in daily conversation we typically are referring to its meaning, not its form (sign). The meaning of symbols emerges from the interpretation process which is triggered by the contextual use of a sign: A sign's relation to forward models, in the pragmatic and syntactic context, produces a meaning - that which is signified. Thus, more than being “stored in a database”, symbols are continuously and dynamically being “computed based on knowledge”. |
Models & Symbols | Both are representations - but models contain more than symbols; if symbols are pointers models are machines. |
So, What Are Models?
Model | A model of something is an information structure that behaves in some ways like the thing being modeled. ‘Model’ here actually means exactly the same as the word when used in the vernacular — look up any dictionary defnition and that is what it means. A model of something is not the thing itself, it is in some way a ‘mirror image’ of it, typically with some unimportant details removed, and represented in a way that allows for various manipulations for the purpose of making predictions (answering questions), where the form of allowed manipulations are particular to the representation of the model and the questions to be answered. |
Example | A model of Earth sits on a shelf in my daughter’s room. With it I can answer questions about the gross layout of continents, and names assigned to various regions as they were around 1977 (because that’s when I got it for my confirmation |
Computational Models | A typical type of question to be answered with computational (mathematical) models are what-if questions, and a typical method of manipulation is running simulations (producing deductions). Along with this we need the appropriate computational machine. |
Model (again) | A 'model' in this conception has a target phenomenon that it applies to, and it has a form of representation, comprehensiveness, and level of detail; these are the primary features that determine what a model is good for. A computational model of the world in raw machine-readable form is not very efficient for quickly identifying all the countries adjacent to Switzerland - for that a traditional globe is much better. |
Model Acquisition | The ability to create models of (observed) phenomena. |
2018©K. R. Thórisson
EOF
/var/www/cadia.ru.is/wiki/data/attic/public/t_720_atai/atai-18/lecture_notes_knowledge_representations.1538844219.txt.gz · Last modified: 2024/04/29 13:33 (external edit)