public:t-720-atai:atai-19:lecture_notes_architectures
Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
public:t-720-atai:atai-19:lecture_notes_architectures [2019/10/01 07:54] – thorisson | public:t-720-atai:atai-19:lecture_notes_architectures [2024/04/29 13:33] (current) – external edit 127.0.0.1 | ||
---|---|---|---|
Line 51: | Line 51: | ||
====Self-Programming ==== | ====Self-Programming ==== | ||
- | [S] State-space search (example: GPS (Newell 1963). The atomic actions are state-changing operators, and a program is represented as a path from the initial state to a final state. Variants of this approach include program search (examples: Gödel Machine (Schmidhuber 2006)): Given the action set A, in principle all programs formed by it can be exhaustively listed and evaluated to find an optimal one according to certain criteria. | ||
- | [P] Production system (example: SOAR (Laird 1987)). Each production rule specifies the condition for a sequence of actions that correspond to a program. Mechanisms that produce new production rules, such as chunking, can be considered self-programming. | ||
- | [R] Reinforcement learning (example: AIXI (Hutter 2007)). When an action of an agent changes the state of the environment, | ||
- | [G] Genetic | + | | What it is | //Self-programming// here means, with respect to some virtual machine < |
+ | | Self-Generated Program | ||
+ | | Historical note | Concept of self-programming is old (J. von Neumann one of the first to talk about self-replication in machines). However, few if any proposals for how to achieve this has been fielded. | ||
+ | | No guarantee | ||
+ | | Why we need it | The inherent limitations of hand-coding methods make traditional manual programming approaches unlikely to reach a level of a human-grade generally intelligent system, simply because to be able to adapt to a wide range of tasks, situations, and domains, a system must be able to modify itself in more fundamental ways than a traditional software system is capable | ||
+ | | Remedy | ||
+ | | The Self of a machine | ||
+ | | Bootstrap code | A.k.a. the " | ||
+ | \\ | ||
+ | \\ | ||
+ | ==== Programming for Self-Programming ==== | ||
+ | | Can we use LISP? | Any language with similar features as LISP (e.g. Haskel, Prolog, etc.), i.e. the ability to inspect itself, turn data into code and code into data, should //in theory// be capable of sustaining a self-programming machine. | ||
+ | | Theory vs. practice | ||
+ | | Why? | Building a machine that can write (sensible, meaningful!) programs means that machine is smart enough to understand the code it produces. If the purpose of its programming is to // | ||
+ | | What can we do? | We must create a programming language with //simple enough// semantics so that a simple machine (perhaps with some clever emergent properties) can use it to bootstrap itself in learning to write programs. | ||
+ | | Does such a language exist? | ||
- | [I] Inductive logic programming (c.f. Muggleton 1994). A program is a statement with a procedural interpretation, | + | \\ |
- | Besides the above fairly well-known AI approaches, we add here two less known ones that have recently become relevant in the context of self-programming: | + | \\ |
- | [E] Evidential reasoning (example: NARS (Wang 2006)). A program is a statement with a procedural interpretation, | + | |
- | [A] Autocatalysis (example: Ikon Flux (Nivel 2007)). In this context the architecture is in large part comprised of a large collection of models, acting as hierarchically organized controllers, | ||
\\ | \\ | ||
Line 72: | Line 82: | ||
====The SOAR Architecture==== | ====The SOAR Architecture==== | ||
- | TBD | + | | What it is | One of the oldest cognitive architectures in history. |
+ | | Why is it important | ||
+ | | How does it work | Reasoning engine does pattern-matching with hand-coded ' | ||
+ | | Recent Additionns | ||
+ | | Missing in Action | ||
+ | |||
+ | |||
+ | |||
+ | SOAR is a relatively mature cognitive architecture that has been used by many researchers worldwide during its 20 year life span. During this time it has also been revised and extended in a number of ways. The architecture consists of heterogenous components that interact during each decision cycle. These are working memory and three types of long-term memory: semantic, procedural and episodic. Working memory is where information related to the present is stored with its contents being supplied by sensors or copied from other memory structures based on relevancy to the present situation. Working memory also contains an activation mechanism, used in conjunction with episodic memory, that indicates the relevancy and usefulness of working memory elements. Production rules are matched and fired on the contents of working memory during the decision cycle, implementing both an associative memory mechanism (as rules can bring data from long-term memory into working memory) and action selection (as rules propose, evaluate and apply operators). Operators are procedural data stored in procedural memory. The application of an operator is carried out by a production rule and either causes changes in the working memory or triggers an external action. In cases where operator selection fails due to insufficient knowledge, an impasse event occurs and a process to resolve the impasse is started. This process involves reasoning and inference upon existing knowledge using the same decision cycle in a recursive fashion, the results of this process are converted to production rules by a process termed chunking. Reinforcement learning is used for production rules relating to operator selection to maximize future rewards in similar situations. One of the most recent additions to the SOAR architecture is sub-symbolic processing used for visual capabilities, | ||
+ | |||
+ | The SOAR architecture provides one of the largest collection of simultaneous running cognitive processes of any cognitive architecture so far. However, there is no explicit mechanism for control of attention and the architecture is not designed for real-time operation. The latter may be especially problematic as execution is in strict step-lock form and in particular, the duration (amount of computation) in each decision cycle can vary greatly due to impasse events that are raised occasionally. One might argue that the development of SOAR has been somewhat characterized by " | ||
+ | |||
+ | There are a few cognitive architectures that somewhat resemble SOAR and can be placed categorically on the same track. These include ICARUS, which has a strong emphasis on embodiment and has shown promise in terms of generality in a number of toy problems such as in-city driving, and LIDA which was developed for the US Navy to automatically organize and negotiate assignments with sailors but does not have embodiment as a design goal. As in SOAR, both of these implement different types of memory in specialized components and have a step-locked decision cycle. | ||
+ | |||
+ | 2013(c)Helgi P. Helgason | ||
\\ | \\ | ||
\\ | \\ | ||
Line 78: | Line 102: | ||
\\ | \\ | ||
+ | ====The AERA System==== | ||
+ | The Auto-catalytic Endogenous Reflective Architecture – AERA – is an AGI-aspiring architectural blueprint that was produced as part of the HUMANOBS FP7 project. It encompasses several fundamentally new ideas in the history of AI, including a new programming language specifically conceived to solve some major limitations of prior efforts in this respect, including self-inspection and self-representation, | ||
+ | |||
+ | AERA's knowledge is stored in models, which essentially encode transformations on input, to produce output. Models have a trigger side (left-hand side) and a result side (right-hand side). In a forward-chaining scenario, when a particular piece of data matches on the left hand of a model (it is only allowed to test the match if the data has high enough saliency and the program has sufficient activation) the model fires, producing the output specified by its left-hand side and injecting it into a global memory store. The semantics of the output is prediction, and the semantics of the input is either fact or prediction. Notice that a model in AERA is not a production rule; a model relating A to B does not mean “A entails B”, it means A predicts B, and it has an associated confidence value. Such models stem invariably (read: most of the time) from the system' | ||
+ | |||
+ | In backward-chaining – to implement the process of abduction – models act the other way around, namely, when some data match the right-hand side, a model produces new data patterned after its left side, whose semantics essentially state that “if you want a B (on the right-hand side) perhaps it would help to get an A (term on the left-hand side). The semantics of either (both the input and output) is “goal”. | ||
+ | |||
+ | A key principle of AERA operation is that of a distributed production process: Each program in AERA has a level of activation that determines if it is allowed to run or not. Every piece of data has a corresponding saliency level that determines how visible it is inside the system. In AERA there is a single memory, but it (typically) embeds groups that allow sets of data and programs to be addressed, e.g. changing their activation, saliency, or existence (via creation or deletion). | ||
+ | |||
+ | \\ | ||
+ | \\ | ||
====High-Level View of AERA==== | ====High-Level View of AERA==== | ||
Line 95: | Line 130: | ||
\\ | \\ | ||
- | ====Autonomous Model Acquisition==== | ||
- | | What it is | The ability to create a model of some target phenomenon // | ||
- | | Challenge | ||
\\ | \\ | ||
\\ | \\ | ||
- | ====Model Acquisition Function==== | ||
- | | {{public: | ||
- | | The agent has a model generation function < | ||
- | | {{public: | ||
- | | A learning agent is situated so as to perceive the effects of the relationships between variables. \\ The agent observes the interaction between the variables for a while, rendering some data about their relations (but not enough to be certain about it, and certainly not enough to create a complete model of it). \\ This generates hypotheses about the relation between variables, in the form of candidate relational models of the observed events. | ||
- | |||
- | \\ | ||
- | \\ | ||
- | ==== Model Generation & Evaluation ==== | ||
- | |||
- | | {{public: | ||
- | | Based on prior observations, | ||
- | | {{public: | ||
- | | The agent' | ||
- | | {{public: | ||
- | | The result of feedback (reinforcement) may result in the deletion, rewriting, or some other modification of the original model selected for prediction. Here the feedback has resulted in a modified model < | ||
- | |||
- | \\ | ||
- | \\ | ||
- | |||
- | |||
- | |||
- | \\ | ||
- | \\ | ||
- | |||
- | |||
- | |||
- | ====Demo Of AERA In Action==== | ||
- | | Demos | The most complex demo of an AERA system was the S1 agent learning to do an interview (in the EU-funded HUMANOBS research project). [[http:// | ||
- | | TV Interview | ||
- | | Data | S1 received realtime timestamped data from the 3D movement of the humans (digitized via appropriate tracking methods at 20 Hz), words generated by a speech recognizer, and prosody (fundamental pitch of voice at 60 Hz, along with timestamped starts and stops). | ||
- | | Seed | The seed consisted of a handful of top-level goals for each agent in the interview (interviewer and interviewee), | ||
- | | What Was Given | * actions: grab, release, point-at, look-at (defined as event types constrained by geometric relationships) \\ * stopping the interview clock ends the session \\ * objects: glass-bottle, | ||
- | | What Had To Be Learned | ||
- | | Result | ||
- | |||
- | |||
- | \\ | ||
- | \\ | ||
- | |||
- | ====The X Architecture==== | ||
- | |||
- | TBD | ||
- | \\ | ||
- | \\ | ||
- | \\ | ||
- | \\ | ||
- | 2018(c)K. R. Thórisson | + | 2019(c)K. R. Thórisson |
\\ | \\ | ||
\\ | \\ | ||
//EOF// | //EOF// |
/var/www/cadia.ru.is/wiki/data/attic/public/t-720-atai/atai-19/lecture_notes_architectures.1569916492.txt.gz · Last modified: 2024/04/29 13:32 (external edit)