| |
public:t_720_atai:atai-18:lecture_notes_knowledge_representations [2018/10/06 16:40] – [Lecture Notes, W9: Knowledge Representation] thorisson | public:t_720_atai:atai-18:lecture_notes_knowledge_representations [2024/04/29 13:33] (current) – external edit 127.0.0.1 |
---|
| Attention | The management of processing, memory, and sensory resources. || | | Attention | The management of processing, memory, and sensory resources. || |
| Meta-Cognition | The ability of a system to reason about itself. || | | Meta-Cognition | The ability of a system to reason about itself. || |
| \\ Understanding | The phenomenon of "understanding" has been neglected in AI and AGI; Modern AI systems do not //understand//. \\ Yet the concept seems crucial when talking about human intelligence; the concept holds explanatory power - we do not assign responsibilities for a task who has a significant lack of understanding of the task. Moreover, the level of understanding can be evaluated. \\ Understanding of a particular phenomenon <m>phi</m> is the potential to perform actions and answer questions with respect to <m>phi</m>. Example: Is an automobile heavier or lighter than a human? || | | \\ Understanding | The phenomenon of "understanding" has been neglected in AI and AGI. Modern AI systems do not //understand//. \\ Yet the concept seems crucial when talking about human intelligence; the concept holds explanatory power - we do not assign responsibilities for a task to someone or something with a demonstrated lack of understanding of the task. Moreover, the level of understanding can be evaluated. \\ Understanding of a particular phenomenon <m>phi</m> is the potential to perform actions and answer questions with respect to <m>phi</m>. Example: Is an automobile heavier or lighter than a human? || |
| | \\ Explanation | When performed by an agent, the ability to transform knowledge about X from a formulation primarily (or only) good for execution with respect to X to a formulation good for being communicated (typically involving some form of linearization, incremental introduction of concepts and issues, in light of an intended receiving agent with a particular a-priori knowledge). \\ Is it possible to explain something that you don't understand? | | | | \\ Explanation | When performed by an agent, the ability to transform knowledge about X from a formulation primarily (or only) good for execution with respect to X to a formulation good for being communicated (typically involving some form of linearization, incremental introduction of concepts and issues, in light of an intended receiving agent with a particular a-priori knowledge). \\ Is it possible to explain something that you don't understand? | |
| \\ Learning | Acquisition of information in a form that enables more successful completion of tasks. We call information in such a form "knowledge" or "practical knowledge". (There is also the concept of "impractical knowledge", which sometimes people feel must be the case of "useless trivia" that seems to be useless for anything, but can in fact turn out to be useful at any point, as for instance using such trivia to wow others with one's knowledge of trivia.) || | | \\ Learning | Acquisition of information in a form that enables more successful completion of tasks. We call information in such a form "knowledge" or "practical knowledge". (There is also the concept of "impractical knowledge", which sometimes people feel must be the case of "useless trivia" that seems to be useless for anything, but can in fact turn out to be useful at any point, as for instance using such trivia to wow others with one's knowledge of trivia.) || |
| Where the Symbols "are" | When we use the term "symbol" in daily conversation we typically are referring to its //meaning//, not its form (sign). The meaning of symbols emerges from the interpretation process which is triggered by the contextual use of a sign: A sign's relation to forward models, in the pragmatic and syntactic context, produces a meaning - that which is //signified//. Thus, more than being "stored in a database", symbols are continuously and dynamically being "computed based on knowledge". | | | Where the Symbols "are" | When we use the term "symbol" in daily conversation we typically are referring to its //meaning//, not its form (sign). The meaning of symbols emerges from the interpretation process which is triggered by the contextual use of a sign: A sign's relation to forward models, in the pragmatic and syntactic context, produces a meaning - that which is //signified//. Thus, more than being "stored in a database", symbols are continuously and dynamically being "computed based on knowledge". | |
| Models & Symbols | Both are representations - but models contain more than symbols; if symbols are pointers models are machines. | | | Models & Symbols | Both are representations - but models contain more than symbols; if symbols are pointers models are machines. | |
| |
| \\ |
| \\ |
| ====So, What Are Models?==== |
| |
| | Model | A model of something is an information structure that behaves in some ways like the thing being modeled. \\ ‘Model’ here actually means exactly the same as the word when used in the vernacular — look up any dictionary defnition and that is what it means. A model of //something// is not the thing itself, it is in some way a ‘mirror image’ of it, typically with some unimportant details removed, and represented in a way that allows for various //manipulations// for the purpose of //making predictions// (//answering questions//), where the form of allowed manipulations are particular to the representation of the model and the questions to be answered. | |
| | Example | A model of Earth sits on a shelf in my daughter’s room. With it I can answer questions about the gross layout of continents, and names assigned to various regions as they were around 1977 (because that’s when I got it for my confirmation :-) ). A model requires a //process for using// it. In this example that process is humans that can read and manipulate smallish objects. | |
| | Computational Models | A typical type of question to be answered with computational (mathematical) models are what-if questions, and a typical method of manipulation is running simulations (producing deductions). Along with this we need the appropriate computational machine. | |
| | \\ Model (again) | A 'model' in this conception has a target phenomenon that it applies to, and it has a form of representation, comprehensiveness, and level of detail; these are the primary features that determine what a model is good for. A computational model of the world in raw machine-readable form is not very efficient for quickly identifying all the countries adjacent to Switzerland - for that a traditional globe is much better. | |
| | Model Acquisition | The ability to create models of (observed) phenomena. | |
| |
| \\ |
| \\ |
| ====System & Architectural Requirements for Using Models==== |
| | Effectiveness | Creation of models must be effective - otherwise a system will spend too much time creating useless or bad models. \\ Making the model creation effective may require e.g. parallelizing the execution of operations on them. | |
| | Efficiency | Operations on models listed above must be efficient lest they interfere with the normal operation of the system / agent. \\ One way to achieve temporal efficiency is to parallelize their execution, and make them simple. | |
| | Scalability | For any moderately interesting / complex environment, a vast number of models may be entertained and considered at any point in time, and thus a large set of //potential// models must be manipulatable by the system / agent. | |
| |
| |
| |
| \\ |
| \\ |
| ==== Problems with Feedback-Only Controllers==== |
| |
| | \\ \\ Thermostat | A cooling thermostat has a built-in supersimple model of its task-environment, one that is sufficient for it to do its job. It consists of a few variables, an on-off switch, two thresholds, and two simple rules that tie these together; the sensed temperature variable, the upper threshold for when to turn the heater on, and the lower threshold for when to turn the heater off. The thermostat never has to decide which model is appropriate, it is “baked into it" by the thermostat’s designer. It is not a predictive (forward) model, this is a strict feedback model. \\ The thermostat cannot change its model, this can only be done by the user opening it and twiddling some thumbscrews. | |
| | Limitation | Because the system designer knows beforehand which signals cause perturbations in <m>o</m> and can hard-wire these from the get-go in the thermostat, there is no motivation to create a model-creating controller (it is much harder!). | |
| | Other "state of the art" systems | The same is true for expert systems, subsumption robots, and general game playing machines: their model is to tightly baked into their architecture by the designer. Yes, there are some variables in these that can be changed automatically “after the machine leaves the lab” (without designer intervention), but they are parameters inside a (more or less) already-determined //model//. | |
| | What Can We Do? | Feed-forward control! Which requires **models**. | |
| |
| |
| \\ |
| \\ |
| |
| ==== Benefits of Combined Feed-forward + Feedback Controllers ==== |
| | Ability to Predict | With the ability to predict comes the ability to deal with events that happen faster than the perception-action cycle of the controller, as well as the ability to anticipate events far into the future. | |
| | \\ Greater Potential to Learn | A machine that is free to create, select, and evaluate models operating on observable and hypothesized variables has potential to learn anything (within the confines of the algorithms it has been given for these operations) because as long as the range of possible models is reasonably broad and general, the topics, tasks, domains, and worlds it could (in theory) handle becomes vastly larger than systems where a particular model is given to the system a priori (I say ‘in theory’ because there are other factors, e.g. the ergodicity of the environment and resource constraints that must be favorable to e.g. the system’s speed of learning). | |
| | Greater Potential for Cognitive Growth | A system that can build models of its own model creation, selection, and evaluation has the ability to improve its own nature. This is in some sense the ultimate AGI (depending on the original blueprint, original seed, and some other factors of course) and therefore we only need two levels of this, in theory, for a self-evolving potentially omniscient/omnipotent (as far as the universe allows) system. | |
| | Bottom Line | AGI without both feed-forward and feed-back mechanisms is fairly unthinkable. | |
| |
| |
\\ | \\ |