public:t-622-arti-08-1:thought_questions
Differences
This shows you the differences between two versions of the page.
Both sides previous revisionPrevious revisionNext revision | Previous revision | ||
public:t-622-arti-08-1:thought_questions [2008/01/06 14:44] – thorisson | public:t-622-arti-08-1:thought_questions [2024/04/29 13:33] (current) – external edit 127.0.0.1 | ||
---|---|---|---|
Line 1: | Line 1: | ||
====== Thought Questions ====== | ====== Thought Questions ====== | ||
+ | |||
+ | Please note that the successful submission of all Thought Question assignments during the course is a prerequisite for taking the final test and thus a prerequisite for finishing the course. | ||
In each lecture where there is new reading material, each student must create one " | In each lecture where there is new reading material, each student must create one " | ||
Line 9: | Line 11: | ||
===== Examples ===== | ===== Examples ===== | ||
+ | ==Chapter 2== | ||
Let's take a look at some example thought questions from chapter 2 (you will NOT hand in thought questions for chapter 2). | Let's take a look at some example thought questions from chapter 2 (you will NOT hand in thought questions for chapter 2). | ||
- | |||
**Examples of good thought questions: | **Examples of good thought questions: | ||
- | Q: Should the critic | + | Q: Should the evaluation |
A: It should be static so that all the experience that has been gathered doesn' | A: It should be static so that all the experience that has been gathered doesn' | ||
A: It should be dynamic because if the environment is dynamic, the critic should be able to adjust to it\\ | A: It should be dynamic because if the environment is dynamic, the critic should be able to adjust to it\\ | ||
Line 26: | Line 28: | ||
A: No, because there is a danger of the agent' | A: No, because there is a danger of the agent' | ||
- | + | **Examples of Bad Thought Questions:**\\ | |
- | **Examples of bad thought questions:**\\ | + | |
Q: What is an agent?\\ | Q: What is an agent?\\ | ||
A: An agent is anything that can be viewed as perceiving its environment through sensors and acting upon that environment through actuators. \\ | A: An agent is anything that can be viewed as perceiving its environment through sensors and acting upon that environment through actuators. \\ | ||
Line 35: | Line 36: | ||
A: Simple reflex, model-based, | A: Simple reflex, model-based, | ||
--This question is bad because it is difficult to see that s/he who wrote it spent any time thinking about the topic. | --This question is bad because it is difficult to see that s/he who wrote it spent any time thinking about the topic. | ||
+ | |||
+ | |||
+ | ==Chapter 24== | ||
+ | |||
+ | From Bjarni Þór Árnason: | ||
+ | |||
+ | **Q**: Imagine a robot with vision, that processes all of its captured data into a 3d world. How can the robot make an educated guess about the back side (side that he doesn' | ||
+ | |||
+ | A: One possibly solution is to invert the object, and assume it's symmetric (in 3d.. is that possible?) | ||
+ | A: An other solution is to use shadows and other objects as a hint. For example, my mousepad is sitting on my desk, and therefore it casts almost no shadow. Therefore it most likely has a solid back that has the same shape as the top side that we see. But it's possible that it's not and is " | ||
+ | |||
+ | **Q**: How can we improve our method of recognizing objects? For example, recognizing a car to be a car, not 4 wheels, an aluminum body, seats, etc. | ||
+ | |||
+ | A: Motion blur (even though it's probably considered noise) could possibly help here. But only if the object is moving. We can see that if the whole object moves at the same speed in the same direction, then it's a pretty safe bet that it's one whole. | ||
+ | An obvious other way (that is most likely how it's solved today) is to take more pictures and compare them picture to picture, but then you have the issue of matching things, that have possibly rotated, turned, moved, etc. | ||
+ | |||
+ | **Q**: The computer vision chapter sets fourth a set of difficult problems with computer vision, such as seperating foreground from background, seperating shadows from physical things, estimating depth, etc. If we were to think outside of the box, how could we possibly solve these problems? | ||
+ | |||
+ | A: As shown in the reading material, image filters such as the SUSAN edge detection filters can be used to some degree. However, I'd like to consider an other possibility... | ||
+ | Using multiple flashes(camera flashes) to create more visible shadows, and elimnating shadows at the same time. | ||
+ | Imagine one flash mounted 1 meter to the right of the camera and other 1 meter to the left(possibly more), and there is an object in front of the camera. | ||
+ | The camera would then take 2 pictures, 1 with each flash, and process them for differences. Using the intersections and unions of the shadows (and the lack of shadows) you should be able to construct an outline of the object. Obviously this method has it's downsides, but it's something that I found worth considering. | ||
+ | This would also give us the advantage of being able to seperate shadows from physical things, as well as seperating foreground from background (flashes have limited range. image could be compared to one taken without flash). | ||
+ | |||
+ | --- | ||
+ | |||
+ | From Brynjar Reynisson: | ||
+ | |||
+ | **Q**: The perception chapter describes various algorithms and mathematical transformations for deciphering visual data. In nature however, all higher level visual processing happens through multilayered neural networks (at least that's what we think). What are some of the pros and cons of trying to go that way for turning raw visual data into categorized information for said data? | ||
+ | |||
+ | Pros: Obviously humans and many animals are very good at dealing with visual changes in the environment (due to such things as lighting and egomotion), also at recognizing things that are only faintly similar as belonging to the same group of things. | ||
+ | How people and animals perceive the world, evolves as the being grows older so learning is continuous. Our traditional way of dealing with neural networks has been to first train them for a period and then use what they' | ||
+ | |||
+ | Cons: We don't know how to do it - what are the functions, what is their relationship, | ||
+ | |||
+ | **Q**: Watching a football game with machine vision on my mind (neurons?), I wonder what it would take to have an agent be able to correctly identity at least some elements that happen in a game? E.g. what team has the ball or is passing it, is the ball in play and if it's out of the field who touched it last (to infer who gets a throw-in or corner kick), to see if the ball crosses the goal line etc. | ||
+ | |||
+ | A: Many aspects of the game seem to be quite easily segmented. If you start with edge detection, you will get the moving players (and referee), the ball, as well as the static painted lines and the goal posts. Following that you could do some sort of coloring analysis for the players to see which team they belong to, or if it's the referee we're talking about. Traditionally, | ||
+ | Depth perception can be tricky, but the static nature of the background (painted lines, goal posts, observer areas) should make the job easier. More difficult jobs include detecting what body parts a player uses to move the ball (everything but the hands is fair game, except if you're talking about the goalkeeper). Another tricky one would be to detect how legal player collision is - but that one is also very tricky for human referees... | ||
+ | |||
\\ | \\ |
/var/www/cadia.ru.is/wiki/data/attic/public/t-622-arti-08-1/thought_questions.1199630688.txt.gz · Last modified: 2024/04/29 13:32 (external edit)