dc.contributor.advisor |
Friedler, Sorelle |
|
dc.contributor.author |
Villalta, Christopher |
|
dc.date.accessioned |
2019-08-21T16:18:33Z |
|
dc.date.available |
2019-08-21T16:18:33Z |
|
dc.date.issued |
2019 |
|
dc.identifier.uri |
http://hdl.handle.net/10066/21628 |
|
dc.description.abstract |
Machine learning has become a popular tool for decision making. As these algorithms continue to make more of our decisions, the question arises of whether they are sound and just. Seeking the answer to this question is a difficult task due to the black box nature of these algorithms, and although researchers are making headway on interpreting currently employed algorithms, many algorithms in machine learning remain uninterpretable. Reinforcement learning is a branch of machine learning that has seen great improvements recently, yet there is a severe lack of interpretability methods for it. In this work we propose an interpretability approach for deep Q-networks by evaluating the influence state experiences have on the network. |
|
dc.description.sponsorship |
Haverford College. Department of Computer Science |
|
dc.language.iso |
eng |
|
dc.rights.uri |
http://creativecommons.org/licenses/by-nc/4.0/ |
|
dc.subject.lcsh |
Reinforcement learning |
|
dc.title |
State Influence Calculations for Deep Q-Networks |
|
dc.type |
Thesis |
|
dc.rights.access |
Haverford users only until 2020-01-01, afterwards Tri-College users only. |
|