TY - GEN
T1 - Reinforcement Learning for Orientation on the Lie Algebra
AU - Alhousani, Naseem
AU - Kose, Hatice
AU - Abu-Dakka, Fares J.
N1 - Publisher Copyright:
© 2023 IEEE.
PY - 2023
Y1 - 2023
N2 - In this paper, we propose a novel framework for Reinforcement Learning on Lie algebra and show how it applies to learning the orientation of the robot's end effector in the task space. The proposed framework is suitable for model-free Reinforcement learning algorithms. Our research is motivated by the fact that in robotics, non-Euclidean data (e.g., orientation) is common in learning manipulation skills, yet neglecting the geometric meaning of such data affects learning performance and accuracy. In particular, our innovation is to apply policy parameterization and learning on the Lie algebra, then map back the learned actions to the hemisphere manifold. The proposed framework opens the door for some model-free Reinforcement learning algorithms designed for Euclidean space to learn non-Euclidean data without change. According to the best of our knowledge, this research work is the first effort in applying a policy parameterization in the context of Reinforcement learning on the Lie algebra of the hemisphere manifold. The results of our experiments provide evidence to support our hypothesis that learning orientation on the Lie algebra is more precise and leads to a superior solution than learning through the normalization of non-Euclidean data.
AB - In this paper, we propose a novel framework for Reinforcement Learning on Lie algebra and show how it applies to learning the orientation of the robot's end effector in the task space. The proposed framework is suitable for model-free Reinforcement learning algorithms. Our research is motivated by the fact that in robotics, non-Euclidean data (e.g., orientation) is common in learning manipulation skills, yet neglecting the geometric meaning of such data affects learning performance and accuracy. In particular, our innovation is to apply policy parameterization and learning on the Lie algebra, then map back the learned actions to the hemisphere manifold. The proposed framework opens the door for some model-free Reinforcement learning algorithms designed for Euclidean space to learn non-Euclidean data without change. According to the best of our knowledge, this research work is the first effort in applying a policy parameterization in the context of Reinforcement learning on the Lie algebra of the hemisphere manifold. The results of our experiments provide evidence to support our hypothesis that learning orientation on the Lie algebra is more precise and leads to a superior solution than learning through the normalization of non-Euclidean data.
KW - Lie algebra
KW - policy optimization
KW - policy search
KW - re-inforcement learning
UR - http://www.scopus.com/inward/record.url?scp=85173489015&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85173489015&partnerID=8YFLogxK
U2 - 10.1109/SIU59756.2023.10224004
DO - 10.1109/SIU59756.2023.10224004
M3 - Conference contribution
AN - SCOPUS:85173489015
T3 - 31st IEEE Conference on Signal Processing and Communications Applications, SIU 2023
BT - 31st IEEE Conference on Signal Processing and Communications Applications, SIU 2023
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 31st IEEE Conference on Signal Processing and Communications Applications, SIU 2023
Y2 - 5 July 2023 through 8 July 2023
ER -