DTC: Deep Tracking Control

  Рет қаралды 16,039

Robotic Systems Lab: Legged Robotics at ETH Zürich

Robotic Systems Lab: Legged Robotics at ETH Zürich

Күн бұрын

We have combined trajectory optimization and reinforcement learning to achieve versatile and robust perceptive legged locomotion.
Published in Science Robotics: www.science.or...
arXiv: doi.org/10.485...
Abstract: Legged locomotion is a complex control problem that requires both accuracy and robustness to cope with real-world challenges. Legged systems have traditionally been controlled using trajectory optimization with inverse dynamics. Such hierarchical model-based methods are appealing due to intuitive cost function tuning, accurate planning, generalization, and most importantly, the insightful understanding gained from more than one decade of extensive research. However, model mismatch and violation of assumptions are common sources of faulty operation. Simulation-based reinforcement learning, on the other hand, results in locomotion policies with unprecedented robustness and recovery skills.
Yet, all learning algorithms struggle with sparse rewards emerging from environments where valid footholds are rare, such as gaps or stepping stones. In this work, we propose a hybrid control architecture that combines the advantages of both worlds to simultaneously achieve greater robustness, foot-placement accuracy, and terrain generalization. Our approach utilizes a model-based planner to roll out a reference motion during training. A deep neural network policy is trained in simulation, aiming to track the optimized footholds. We evaluate the accuracy of our locomotion pipeline on sparse terrains, where pure data-driven methods are prone to fail. Furthermore, we demonstrate superior robustness in the presence of slippery or deformable ground when compared to model-based counterparts. Finally, we show that our proposed tracking controller generalizes across different trajectory optimization methods not seen during training. In conclusion, our work unites the predictive capabilities and optimality guarantees of online planning with the inherent robustness attributed to offline learning.
Authors: Fabian Jenelten, Junzhe He, Farbod Farshidian, and Marco Hutter
Video: Fabian Jenelten
#ANYmal #leggedrobot #robot #robotics #robotdog #AI #reinforcementlearning #rl #rescue #innovation #armasuisse #arche2023 #scienceresearch #stepping

Пікірлер: 38
@TheCrassEnnui_2501
@TheCrassEnnui_2501 8 ай бұрын
To say this is impressive would be an understatement. I look forward to what comes next and wonder what the state of the art will be in 10 years.
@leggedrobotics
@leggedrobotics 8 ай бұрын
Thanks a lot. We hope that in 10 years these robots can be deployed fully autonomously in hazardous real-world environments to assist humans or even completely remove them from harms way.
@gloudsdu
@gloudsdu 8 ай бұрын
The most natural movement I've ever seen, so good
@avinashthakur80
@avinashthakur80 8 ай бұрын
Truly impressive. I had previously tried to implement a walking gait for walking robot, but the MPC controller never worked out of the simulation. Seeing this paper motivates me to try it again.
@leggedrobotics
@leggedrobotics 8 ай бұрын
We are very happy to hear that we can inspire people all round the world. No control methods works out of the box. There is always a lot of tuning and engineering work required to successfully bridge the sim2real gap.
@divinusfilius3771
@divinusfilius3771 6 ай бұрын
This is way more impressive than ANY other demo I've seen lately from anyone. Just don't go greedy and have governments take up all your contracts.
@BalajiSankar
@BalajiSankar 8 ай бұрын
Your work on walking in the soft deformable floor when the camera perceives a rigid surface is amazing. I am floored.
@c016smith52
@c016smith52 8 ай бұрын
Great video and impressive work! Can't wait to see more about swarm planning and coordination in the future, thanks for sharing.
@leggedrobotics
@leggedrobotics 8 ай бұрын
These are indeed very exciting research topics. Check our channel for updates about coordinated robot collaboration.
@deTobs
@deTobs 8 ай бұрын
Absolutely amazing!
@AnianBrosig-yv3jn
@AnianBrosig-yv3jn 8 ай бұрын
pretty amazing work
@Cian-_-
@Cian-_- 8 ай бұрын
that was awesome!
@Skora420
@Skora420 8 ай бұрын
Do these robots have some kind of pressure sensors on their limbs? It seems to me that this makes it extremely easier to navigate on soft and uncertain ground. Although the design is not perfect yet, it moves incredibly well on a hard surface and I am sure it will be very useful in helping people.
@leggedrobotics
@leggedrobotics 8 ай бұрын
They do not have force sensors on their feet because the repeated impacts will most likely break them. However, we can measure the torque at the actuators, which allows to estimate the interaction force. But you are probably right: if the robot would also have a sense of touch, it would understand its surroundings even better.
@waynephillips2613
@waynephillips2613 8 ай бұрын
They are ahead of many other robotics companies including Tesla when it comes to locomotion...but what about autonomy? The robot needs to do this autonomously for this to be impressive. And self charge/self sustain
@leggedrobotics
@leggedrobotics 8 ай бұрын
Absolutely agree! Navigation is the next big step that we try to tackle.
@deepanshu5592
@deepanshu5592 8 ай бұрын
This is so awesome
@matthiaskubisch
@matthiaskubisch 8 ай бұрын
chapeau ❤
@divinusfilius3771
@divinusfilius3771 6 ай бұрын
So with artificial training environments with realistic physics you can literally train this thing fast forwarded, give it thousands of years of experience in days or something?? That's so cool. We're all fked.
@omnianti0
@omnianti0 8 ай бұрын
it start to look serious atill wonder the autonomy of such battery
@divinusfilius3771
@divinusfilius3771 6 ай бұрын
Really incredible work. Wow
@Suushidesu
@Suushidesu 8 ай бұрын
It stumbles very "environment aware" looks like an actuall creature :o
@VK-qh6pr
@VK-qh6pr 8 ай бұрын
Apart from tge very impressive results can you say ehat you have used to simulate/train the trajectory logic?
@leggedrobotics
@leggedrobotics 8 ай бұрын
The reference trajectories come from a model-based optimization method. In our case we have used TAMOLS (arxiv.org/abs/2206.14049), but it could be substituted with any other method aswell.
@NeuroScientician
@NeuroScientician 8 ай бұрын
Is this open-source project?
@leggedrobotics
@leggedrobotics 8 ай бұрын
The TO method, unfortunately, is not open.source because it depends on ANYmal research. We currently considering to open-source the RL part independently.
@Ree1981
@Ree1981 8 ай бұрын
I feel like this will soon be child's play for anyone, with how fast AI is improving. Model in the slipperiness and sliding footholds in the 3D simulation and train a new AI from the ground up just to see if it works. I believe we've learned from other new AI systems that "human intervention", like how these robots always 'step' rhythmically, even when 'standing still', is something that holds them back in the end. I don't know what type of AI system this bot is using, but it definitely seems like an advanced iteration of a fairly old way of doing things in the AI space. Don't be afraid to step out of your comfort zone and have a completely separate AI try to learn things solely from 3D environments, with little to no 'humans thinks it should be done this way' intervention.
@leggedrobotics
@leggedrobotics 8 ай бұрын
In this work, we have combined model-based with model-free control. While this does combine the advantages of both methods, it also merges some of their disadvantages, one of them being the fixed gait. This assumption reduces the number of potential motions but also ensures that the learned behaviors transfer well to reality.
@Ree1981
@Ree1981 8 ай бұрын
@@leggedrobotics Fair, and I am just speculating. But let me ask, is this the same movement system as other popular 'dogs', like Spot? If no, what are they (Boston Dynamics) doing differently? Because their robots definitely seem... well, more agile to be honest. Haven't seen Spot in a while though.
@leggedrobotics
@leggedrobotics 8 ай бұрын
@@Ree1981 Most quadrupedal robots, including Spot, are deployed with a gait pattern called "trot". This type of foot alternation can typically be observed with horses, zebras, and deers when they target intermediate fast locomotion velocities. It is a very stable gait that does not requite large lateral or vertical base motions. Because of these properties, it also emerges with learning-based methods, even when not specifically enforced (in our case, however, we do enforce it). Spot's legs are are much lighter and longer than those of ANYmal. This allows for fast gait cycles and large steps, without moving a large amount of intertia. Think, for example, about a pitbull (ANYmal) and a gazell (Spot). Both animals have specialized locomotion behaviors based on their kinematic structure and dynamic properties.
@Ree1981
@Ree1981 8 ай бұрын
@@leggedrobotics Ok, so it could be that my opinion was spot-on (pun intended). Human intervention could be slowing your progress. Because let's face it, ANYmal looks slow and clunky by today's standards. Consider a new approach.
@averagehummus
@averagehummus Ай бұрын
Black mirror vibes
@divinusfilius3771
@divinusfilius3771 6 ай бұрын
I can see IRS ordering 200,000, oh these will find you even under your house. Attach some lasers to the front, you WILL pay your taxes.
@tingdongyan7502
@tingdongyan7502 8 ай бұрын
black mirror is coming
@leggedrobotics
@leggedrobotics 8 ай бұрын
Our robots don't harm humans, animals, or other robots ;)
@peterlustig2048
@peterlustig2048 8 ай бұрын
@@leggedrobotics A robot may not injure a human being or, through inaction, allow a human being to come to harm. :)
@gf2e
@gf2e 6 ай бұрын
@@peterlustig2048Based on that rule, why doesn’t ANYmal try and grab a cigarette from any smoking human it encounters? :)
@peterlustig2048
@peterlustig2048 6 ай бұрын
@@gf2e lol, love it! The second law states: A robot must obey the orders given it by human beings except where such orders would conflict with the First Law. So I guess you still have a vaild point. It is really hard to state what we want robots to do :)
DeepGait: Planning and Control of Quadrupedal Gaits using Deep Reinforcement Learning (Presentation)
8:18
Robotic Systems Lab: Legged Robotics at ETH Zürich
Рет қаралды 20 М.
Nature's Incredible ROTATING MOTOR (It’s Electric!) - Smarter Every Day 300
29:37
小路飞嫁祸姐姐搞破坏 #路飞#海贼王
00:45
路飞与唐舞桐
Рет қаралды 24 МЛН
My Daughter's Dumplings Are Filled With Coins #funny #cute #comedy
00:18
Funny daughter's daily life
Рет қаралды 17 МЛН
Win This Dodgeball Game or DIE…
00:36
Alan Chikin Chow
Рет қаралды 42 МЛН
Alternative to bearings for tiny robots
12:05
Breaking Taps
Рет қаралды 793 М.
This new type of illusion is really hard to make
17:58
Steve Mould
Рет қаралды 595 М.
DIY Laser Image Projector (100ft+ Range!)
20:08
Ben Makes Everything
Рет қаралды 329 М.
Understanding Porsche's New Six Stroke Engine Patent
21:57
driving 4 answers
Рет қаралды 1,7 МЛН
ANYmal Parkour: Learning Agile Navigation for Quadrupedal Robots
4:52
Robotic Systems Lab: Legged Robotics at ETH Zürich
Рет қаралды 22 М.
Does our robot have bird legs?
4:21
Agility Robotics
Рет қаралды 52 М.
Model Based RL Finally Works!
28:01
Edan Meyer
Рет қаралды 34 М.
Reinforcement Learning: Machine Learning Meets Control Theory
26:03
Steve Brunton
Рет қаралды 274 М.
A hybrid control architecture that combines advantages of current quadruped robot controls
10:46
Science X: Phys.org, Medical Xpress, Tech Xplore
Рет қаралды 1,2 М.
Tutorial: Gait and Trajectory Optimization for Legged Robots
28:41
Robotic Systems Lab: Legged Robotics at ETH Zürich
Рет қаралды 38 М.
Evolution of PhoneVision
0:18
PhoneVision
Рет қаралды 1,7 МЛН
3x 2x 1x 0.5x 0.3x... #iphone
0:10
Aksel Alze
Рет қаралды 2,8 МЛН
Hardware tools repair tool high performance tool
0:16
Hardware tools
Рет қаралды 3,5 МЛН
Китайцы сделали телефон БАЯН
0:14
Собиратель новостей
Рет қаралды 1,6 МЛН
Cómo pensaba que sería el iPhone 16 😭
0:12
Alan
Рет қаралды 2 МЛН