Not Only Rewards But Also Constraints: Applications on Legged Robot Locomotion

  Рет қаралды 7,038

RaiLab Kaist

RaiLab Kaist

10 ай бұрын

arxiv.org/abs/2308.12517

Пікірлер: 9
@TextZip
@TextZip 9 ай бұрын
Hi, the video and the paper are really impressive. I wanted to know what simulation platform was used and if the code will be made public..?
@railabkaist9016
@railabkaist9016 9 ай бұрын
You can download the simulator at: raisim.com/ The code will be made public after the paper is accepted
@snuffybox
@snuffybox 10 ай бұрын
You should link the paper
@NowayJose14
@NowayJose14 9 ай бұрын
It's in the description, chief
@snuffybox
@snuffybox 9 ай бұрын
@@NowayJose14 it wasn't when I commented
@user-le3uf8hw5d
@user-le3uf8hw5d 10 ай бұрын
Hello, I hope you're well. While my question isn't directly related to the paper, may I inquire about it? I'm wondering how significant the reality-gap is. Can you consistently expect that every learned policies can be transferred seamlessly to the actual robot? Additionally, would the behavior of both the simulation and the real-world robot be nearly identical? Lastly, if one were to forgo the teacher-student structure, might there be a noticeable decrease in performance?
@user-pg6ym1th9
@user-pg6ym1th9 9 ай бұрын
Thank you for your interest in our work. What we wanted to claim in this work is to use both rewards and constraints when designing learning-based controllers for complex robotic systems. We used the teacher-student learning framework but it is not limited to it. Other methods such as vanilla learning [1] or asymmetric learning [2] can also be used. In our experience, the sim-to-real gap highly depends on the characteristics of the robotic system you are working on (e.g., actuator mechanism, system software latency, actuation bandwidth) rather than the learning algorithm itself. Based on the characteristics of your system, you should select appropriate methods to solve the sim-to-real gap (e.g., domain randomization, domain adaptation, actuator networks [1]). In our case (Raibo, Mini-cheetah), domain randomization was enough. [1] Hwangbo, Jemin, et al. "Learning agile and dynamic motor skills for legged robots." Science Robotics 4.26 (2019): eaau5872. [2] Nahrendra, I. Made Aswin, Byeongho Yu, and Hyun Myung. "Dreamwaq: Learning robust quadrupedal locomotion with implicit terrain imagination via deep reinforcement learning." 2023 IEEE International Conference on Robotics and Automation (ICRA). IEEE, 2023.
@user-le3uf8hw5d
@user-le3uf8hw5d 9 ай бұрын
@@user-pg6ym1th9 Thanks for the detailed discussion! I'm working on a humanoid RL and was wondering if there could be a way to further reduce the reality gap. Thanks for your kind explanation. :)
@marshallmcluhan33
@marshallmcluhan33 10 ай бұрын
Oh thank god they don't just care about collecting more meaningless tokens.
DTC: Deep Tracking Control
6:46
Robotic Systems Lab: Legged Robotics at ETH Zürich
Рет қаралды 14 М.
Talk 09: L. Marchioni, PAL Robotics | ICRA2021 WS on Legged Robots
26:11
БОЛЬШОЙ ПЕТУШОК #shorts
00:21
Паша Осадчий
Рет қаралды 10 МЛН
아이스크림으로 체감되는 요즘 물가
00:16
진영민yeongmin
Рет қаралды 56 МЛН
Learning quadrupedal locomotion on deformable terrain
5:04
RaiLab Kaist
Рет қаралды 9 М.
The moment we stopped understanding AI [AlexNet]
17:38
Welch Labs
Рет қаралды 636 М.
Stepping Up | Reinforcement Learning with Spot | Boston Dynamics
5:29
Boston Dynamics
Рет қаралды 182 М.
SLAM Robot Mapping - Computerphile
11:35
Computerphile
Рет қаралды 126 М.
DeepGait: Planning and Control of Quadrupedal Gaits using Deep Reinforcement Learning (Presentation)
8:18
Robotic Systems Lab: Legged Robotics at ETH Zürich
Рет қаралды 20 М.
We Need to Rethink Exercise - The Workout Paradox
12:00
Kurzgesagt – In a Nutshell
Рет қаралды 385 М.
Robot Dog Learns to Walk - Bittle Reinforcement Learning p.3
22:06
Vision-free MIT Cheetah
2:49
Massachusetts Institute of Technology (MIT)
Рет қаралды 3,8 МЛН
БОЛЬШОЙ ПЕТУШОК #shorts
00:21
Паша Осадчий
Рет қаралды 10 МЛН