Title:
Learning from Observation Using Primitives

dc.contributor.advisor Atkeson, Christopher G.
dc.contributor.author Bentivegna, Darrin Charles en_US
dc.contributor.committeeMember Cheng, Gordon
dc.contributor.committeeMember Hodgins, Jessica K.
dc.contributor.committeeMember Koenig, Sven
dc.contributor.committeeMember Balch, Tucker
dc.contributor.department Computing en_US
dc.date.accessioned 2005-03-02T22:44:05Z
dc.date.available 2005-03-02T22:44:05Z
dc.date.issued 2004-07-13 en_US
dc.description.abstract Learning without any prior knowledge in environments that contain large or continuous state spaces is a daunting task. For robots that operate in the real world, learning must occur in a reasonable amount of time. Providing a robot with domain knowledge and also with the ability to learn from watching others can greatly increase its learning rate. This research explores learning algorithms that can learn quickly and make the most use of information obtained from observing others. Domain knowledge is encoded in the form of primitives, small parts of a task that are executed many times while a task is being performed. This thesis explores and presents many challenges involved in programming robots to learn and adapt to environments that humans operate in. A "Learning from Observation Using Primitives" framework has been created that provides the means to observe primitives as they are performed by others. This information is used by the robot in a three level process as it performs in the environment. In the first level the robot chooses a primitive to use for the observed state. The second level decides on the manner in which the chosen primitive will be performed. This information is then used in the third level to control the robot as necessary to perform the desired action. The framework also provides a means for the robot to observe and evaluate its own actions as it performs in the environment which allows the robot to increase its performance of selecting and performing the primitives. The framework and algorithms have been evaluated on two testbeds: Air Hockey and Marble Maze. The tasks are done both by actual robots and in simulation. Our robots have the ability to observe humans as they operate in these environments. The software version of Air Hockey allows a human to play against a cyber player and the hardware version allows the human to play against a 30 degree-of-freedom humanoid robot. The implementation of our learning system in these tasks helps to clearly present many issues involved in having robots learn and perform in dynamic environments. en_US
dc.description.degree Ph.D. en_US
dc.format.extent 46192366 bytes
dc.format.extent 7360965 bytes
dc.format.extent 40344154 bytes
dc.format.extent 92278888 bytes
dc.format.mimetype application/octet-stream
dc.format.mimetype application/pdf
dc.format.mimetype application/octet-stream
dc.format.mimetype application/octet-stream
dc.identifier.uri http://hdl.handle.net/1853/5100
dc.language.iso en_US
dc.publisher Georgia Institute of Technology en_US
dc.subject Locally weighted learning
dc.subject Reinforcement learning
dc.subject Imitation en_US
dc.title Learning from Observation Using Primitives en_US
dc.type Text
dc.type.genre Dissertation
dspace.entity.type Publication
local.contributor.corporatename College of Computing
relation.isOrgUnitOfPublication c8892b3c-8db6-4b7b-a33a-1b67f7db2021
Files
Original bundle
Now showing 1 - 4 of 4
No Thumbnail Available
Name:
bentivegna_darrin_c_200407_maze.avi
Size:
44.05 MB
Format:
Unknown data format
Description:
Thumbnail Image
Name:
bentivegna_darrin_c_200407_phd.pdf
Size:
7.02 MB
Format:
Adobe Portable Document Format
Description:
No Thumbnail Available
Name:
bentivegna_darrin_c_200407_airhockey_1.avi
Size:
38.48 MB
Format:
Unknown data format
Description:
No Thumbnail Available
Name:
bentivegna_darrin_c_200407_airhockey_2.avi
Size:
88 MB
Format:
Unknown data format
Description: