Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
Low dimensional state representation learning with reward-shaped priors
Robotics and Mechatronics University of Twente, Enschede, Netherlands.
Robotics and Mechatronics University of Twente, Enschede, Netherlands.
Robotics and Mechatronics University of Twente, Enschede, Netherlands.
Datamanagement and Biometrics, University of Twente, Enschede, Netherlands.
Show others and affiliations
2020 (English)In: Proceedings - International Conference on Pattern Recognition, Institute of Electrical and Electronics Engineers (IEEE), 2020, p. 3736-3743Conference paper, Published paper (Refereed)
Abstract [en]

Reinforcement Learning has been able to solve many complicated robotics tasks without any need for feature engineering in an end-to-end fashion. However, learning the optimal policy directly from the sensory inputs, i.e the observations, often requires processing and storage of a huge amount of data. In the context of robotics, the cost of data from real robotics hardware is usually very high, thus solutions that achieve high sample-efficiency are needed. We propose a method that aims at learning a mapping from the observations into a lower-dimensional state space. This mapping is learned with unsupervised learning using loss functions shaped to incorporate prior knowledge of the environment and the task. Using the samples from the state space, the optimal policy is quickly and efficiently learned. We test the method on several mobile robot navigation tasks in a simulation environment and also on a real robot. A video of our experiments can be found at: https://youtu.be/dgWxmfSv95U.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE), 2020. p. 3736-3743
Keywords [en]
Reinforcement Learning, Robotics, State Representation Learning, Digital storage, Mapping, Mobile robots, Pattern recognition, Feature engineerings, Loss functions, Low dimensional, Mobile Robot Navigation, Optimal policies, Prior knowledge, Simulation environment, State representation
National Category
Robotics
Identifiers
URN: urn:nbn:se:hj:diva-54224DOI: 10.1109/ICPR48806.2021.9412421ISI: 000678409203111Scopus ID: 2-s2.0-85110530994ISBN: 9781728188089 (print)OAI: oai:DiVA.org:hj-54224DiVA, id: diva2:1584810
Conference
25th International Conference on Pattern Recognition, ICPR 2020, 10 January 2021 through 15 January 2021
Available from: 2021-08-13 Created: 2021-08-13 Last updated: 2021-08-26Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Sirmacek, Beril

Search in DiVA

By author/editor
Sirmacek, Beril
By organisation
JTH, Department of Computer Science and InformaticsJönköping AI Lab (JAIL)
Robotics

Search outside of DiVA

GoogleGoogle Scholar

doi
isbn
urn-nbn

Altmetric score

doi
isbn
urn-nbn
Total: 125 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf