Detailed Information

Cited 4 time in webofscience Cited 7 time in scopus
Metadata Downloads

StARformer: Transformer With State-Action-Reward Representations for Robot Learning

Full metadata record
DC Field Value Language
dc.contributor.authorShang, Jinghuan-
dc.contributor.authorLi, Xiang-
dc.contributor.authorKahatapitiya, Kumara-
dc.contributor.authorLee, Yu-Cheol-
dc.contributor.authorRyoo, Michael S.-
dc.date.accessioned2024-08-08T08:31:01Z-
dc.date.available2024-08-08T08:31:01Z-
dc.date.issued2023-11-
dc.identifier.issn0162-8828-
dc.identifier.issn1939-3539-
dc.identifier.urihttps://scholarworks.dongguk.edu/handle/sw.dongguk/20484-
dc.description.abstractReinforcement Learning (RL) can be considered as a sequence modeling task, where an agent employs a sequence of past state-action-reward experiences to predict a sequence of future actions. In this work, we propose State-Action-Reward Transformer (StARformer), a Transformer architecture for robot learning with image inputs, which explicitly models short-term state-action-reward representations (StAR-representations), essentially introducing a Markovian-like inductive bias to improve long-term modeling. StARformer first extracts StAR-representations using self-attending patches of image states, action, and reward tokens within a short temporal window. These StAR-representations are combined with pure image state representations, extracted as convolutional features, to perform self-attention over the whole sequence. Our experimental results show that StARformer outperforms the state-of-the-art Transformer-based method on image-based Atari and DeepMind Control Suite benchmarks, under both offline-RL and imitation learning settings. We find that models can benefit from our combination of patch-wise and convolutional image embeddings. StARformer is also more compliant with longer sequences of inputs than the baseline method. Finally, we demonstrate how StARformer can be successfully applied to a real-world robot imitation learning setting via a human-following task. © 1979-2012 IEEE.-
dc.format.extent16-
dc.language영어-
dc.language.isoENG-
dc.publisherIEEE-
dc.titleStARformer: Transformer With State-Action-Reward Representations for Robot Learning-
dc.typeArticle-
dc.publisher.location미국-
dc.identifier.doi10.1109/TPAMI.2022.3204708-
dc.identifier.scopusid2-s2.0-85137939771-
dc.identifier.wosid001085050900010-
dc.identifier.bibliographicCitationIEEE Transactions on Pattern Analysis and Machine Intelligence, v.45, no.11, pp 12862 - 12877-
dc.citation.titleIEEE Transactions on Pattern Analysis and Machine Intelligence-
dc.citation.volume45-
dc.citation.number11-
dc.citation.startPage12862-
dc.citation.endPage12877-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalWebOfScienceCategoryComputer Science, Artificial Intelligence-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.subject.keywordAuthorimitation learning-
dc.subject.keywordAuthorreinforcement learning-
dc.subject.keywordAuthorrobot learning-
dc.subject.keywordAuthorTransformer-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Engineering > Department of Information and Communication Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Related Researcher

Researcher Lee, Yu Cheol photo

Lee, Yu Cheol
College of Engineering (Department of Information and Communication Engineering)
Read more

Altmetrics

Total Views & Downloads

BROWSE