Studies in Computational Intelligence, volume 925 SCI, pages 212-221
Learning Embodied Agents with Policy Gradients to Navigate in Realistic Environments
Staroverov Alexey
1
,
Vetlin Vladislav
2
,
Makarenko Stepan
1
,
Naumov Anton
2
,
Panov Aleksandr I
1, 3
Publication type: Book Chapter
Publication date: 2020-10-02
Quartile SCImago
Q4
Quartile WOS
—
Impact factor: —
ISSN: 1860949X, 18609503
Abstract
Indoor navigation is one of the main tasks in robotic systems. Most decisions in this area rely on ideal agent coordinates and a pre-known room map. However, the high accuracy of indoor localization cannot be achieved in realistic scenarios. For example, the GPS has low accuracy in the room; odometry often gives much noise for accurate positioning, etc. In this paper, we conducted a study of the navigation problem in the realistic Habitat simulator. We proposed a method based on the neural network approach and reinforcement learning that takes into account these factors. The most promising recent approaches were DDPPO and ANM for agent control and DF-VO for localization, during the analysis of which a new approach was developed. This method takes into account the non-determinism of the robot’s actions and the noise level of data from its sensors.
Citations by journals
1
2
|
|
Lecture Notes in Networks and Systems
|
Lecture Notes in Networks and Systems
2 publications, 100%
|
1
2
|
Citations by publishers
1
2
|
|
Springer Nature
|
Springer Nature
2 publications, 100%
|
1
2
|
- We do not take into account publications that without a DOI.
- Statistics recalculated only for publications connected to researchers, organizations and labs registered on the platform.
- Statistics recalculated weekly.
{"yearsCitations":{"type":"bar","data":{"show":true,"labels":[2021,2022],"ids":[0,0],"codes":[0,0],"imageUrls":["",""],"datasets":[{"label":"Citations number","data":[1,1],"backgroundColor":["#3B82F6","#3B82F6"],"percentage":["50","50"],"barThickness":null}]},"options":{"indexAxis":"x","maintainAspectRatio":true,"scales":{"y":{"ticks":{"precision":0,"autoSkip":false,"font":{"family":"Montserrat"},"color":"#000000"}},"x":{"ticks":{"stepSize":1,"precision":0,"font":{"family":"Montserrat"},"color":"#000000"}}},"plugins":{"legend":{"position":"top","labels":{"font":{"family":"Montserrat"},"color":"#000000"}},"title":{"display":true,"text":"Citations per year","font":{"size":24,"family":"Montserrat","weight":600},"color":"#000000"}}}},"journals":{"type":"bar","data":{"show":true,"labels":["Lecture Notes in Networks and Systems"],"ids":[17269],"codes":[0],"imageUrls":["\/storage\/images\/resized\/voXLqlsvTwv5p3iMQ8Dhs95nqB4AXOG7Taj7G4ra_medium.webp"],"datasets":[{"label":"","data":[2],"backgroundColor":["#3B82F6"],"percentage":[100],"barThickness":13}]},"options":{"indexAxis":"y","maintainAspectRatio":false,"scales":{"y":{"ticks":{"precision":0,"autoSkip":false,"font":{"family":"Montserrat"},"color":"#000000"}},"x":{"ticks":{"stepSize":null,"precision":0,"font":{"family":"Montserrat"},"color":"#000000"}}},"plugins":{"legend":{"position":"top","labels":{"font":{"family":"Montserrat"},"color":"#000000"}},"title":{"display":true,"text":"Journals","font":{"size":24,"family":"Montserrat","weight":600},"color":"#000000"}}}},"publishers":{"type":"bar","data":{"show":true,"labels":["Springer Nature"],"ids":[8],"codes":[0],"imageUrls":["\/storage\/images\/resized\/voXLqlsvTwv5p3iMQ8Dhs95nqB4AXOG7Taj7G4ra_medium.webp"],"datasets":[{"label":"","data":[2],"backgroundColor":["#3B82F6"],"percentage":[100],"barThickness":13}]},"options":{"indexAxis":"y","maintainAspectRatio":false,"scales":{"y":{"ticks":{"precision":0,"autoSkip":false,"font":{"family":"Montserrat"},"color":"#000000"}},"x":{"ticks":{"stepSize":null,"precision":0,"font":{"family":"Montserrat"},"color":"#000000"}}},"plugins":{"legend":{"position":"top","labels":{"font":{"family":"Montserrat"},"color":"#000000"}},"title":{"display":true,"text":"Publishers","font":{"size":24,"family":"Montserrat","weight":600},"color":"#000000"}}}}}
Metrics
Cite this
GOST |
RIS |
BibTex
Cite this
GOST
Copy
Staroverov A. et al. Learning Embodied Agents with Policy Gradients to Navigate in Realistic Environments // Studies in Computational Intelligence. 2020. Vol. 925 SCI. pp. 212-221.
GOST all authors (up to 50)
Copy
Staroverov A., Vetlin V., Makarenko S., Naumov A., Panov A. I. Learning Embodied Agents with Policy Gradients to Navigate in Realistic Environments // Studies in Computational Intelligence. 2020. Vol. 925 SCI. pp. 212-221.
Cite this
RIS
Copy
TY - GENERIC
DO - 10.1007/978-3-030-60577-3_24
UR - https://doi.org/10.1007%2F978-3-030-60577-3_24
TI - Learning Embodied Agents with Policy Gradients to Navigate in Realistic Environments
T2 - Studies in Computational Intelligence
AU - Staroverov, Alexey
AU - Vetlin, Vladislav
AU - Makarenko, Stepan
AU - Naumov, Anton
AU - Panov, Aleksandr I
PY - 2020
DA - 2020/10/02 00:00:00
PB - Springer Nature
SP - 212-221
VL - 925 SCI
SN - 1860-949X
SN - 1860-9503
ER -
Cite this
BibTex
Copy
@incollection{2020_Staroverov,
author = {Alexey Staroverov and Vladislav Vetlin and Stepan Makarenko and Anton Naumov and Aleksandr I Panov},
title = {Learning Embodied Agents with Policy Gradients to Navigate in Realistic Environments},
publisher = {Springer Nature},
year = {2020},
volume = {925 SCI},
pages = {212--221},
month = {oct}
}
Profiles