Control of a Quadrotor with Reinforcement Learning

  Рет қаралды 44,428

aslteam

aslteam

Күн бұрын

Пікірлер: 38
@praveengudhi2247
@praveengudhi2247 7 жыл бұрын
You guys are amazing, thank you for showing the direction
@trancongnguyen1252
@trancongnguyen1252 4 жыл бұрын
Thank you very much, that is really awesome to see
@niharranjanpradhan169
@niharranjanpradhan169 7 жыл бұрын
Sir! Which simulator have u used for training?
@yuanyema5578
@yuanyema5578 5 жыл бұрын
The performance of the quadrotor is excellent. But i have a question about the reinforcement learning. I want to know what a role the RL algorithm play in the control system? navigation or attitude control? RL algorithm just provide where to move, and PID controler to finish attitude control? Or both attitude control and location control are all RL algorithm`s duty? Thanks for your response.
@Inviaz
@Inviaz 5 жыл бұрын
These 4 PWM actions can be perform at the same time in one iteration loop?
@Edin12n
@Edin12n 5 жыл бұрын
Hello, What a great video. I'm new to the subject of Reinforced Learning and hoped I could ask a question. Here goes: Does the ability of the drone to recover depend on input from the various sensors e.g. gyro. So say the drone flew from shade to sunlight, would there need to be a temperature sensor on board to allow it to cope with any sudden movement associated with moving from hot to cold. Or does the reinforced learning model not care about any of that and just learn to deal with whatever it encounters (so you could throw away the temperature sensor) and it would stabilize just fine with any sudden movement associated with hot to cold. Thanks
@manukyananush88
@manukyananush88 6 жыл бұрын
I've read the paper however I could not understand how do you define the 4 actions? Each action is each rotor's velocity, right? But then how do they choose the velocity in the beginning?
@emmanuelezenwere
@emmanuelezenwere 6 жыл бұрын
Hello Anush. Have you made any progress with implementing the models from the paper?
@niharranjanpradhan169
@niharranjanpradhan169 5 жыл бұрын
As explained in the paper , there are two networks, one of them is a policy network. The policy network gives the PWM or rpm needed for each motor.
@srinathtangudu4899
@srinathtangudu4899 5 жыл бұрын
what drone did you use? i planning to buy a drone for my research to do similar things. can some suggest me a drone
@keashavnair6772
@keashavnair6772 4 жыл бұрын
The answer is in the V.Experiments section of the paper.
@brucebaxter6923
@brucebaxter6923 3 жыл бұрын
Build your own
@kuan-holao4561
@kuan-holao4561 7 жыл бұрын
Cool work! I just completed a similar project, but I used DDPG with SNNs, nice to see implementation on real quadcopter, have you tried different tasks?
@muhammadharis-nr8yn
@muhammadharis-nr8yn 4 жыл бұрын
can you tell me how did you simulate your project. can you provide me the simulation code and model in matlab?
@ebert7955
@ebert7955 6 жыл бұрын
Random question here: I'm doing something similar using PPO on a quadrotor (simple simulation using OpenAI gym). I'm trying to get the cable suspended load case now, but still struggling with the reward function. Terminal states are way more difficult to handle than the standard quad.
@cypreessDK
@cypreessDK 7 жыл бұрын
What simulation engine did you used? Was that Unity? How did you dealed with reality-gap - what was your approach here. Did the copter flight on policy trained only in simulation?
@InkyuSa
@InkyuSa 7 жыл бұрын
Please have a look our repo and corresponding paper mentioned in the video description. A short answer for "Did the copter flight on policy trained only in simulation?" is yes, simulation only. Thanks for your interests.
@arthurmaffre1454
@arthurmaffre1454 4 жыл бұрын
Hello good job, excuse me I'm a beginner and I don't understand why the value network is essential for learning. With only one value of distance between the quad and the point this could not be enough.
@pratikprajapati8620
@pratikprajapati8620 4 жыл бұрын
This is great indeed
@yugalsharma13
@yugalsharma13 2 жыл бұрын
Amazing work! How did you guys handle the noise in sensors (Gyro/Accelerometer) ?
@harshavardhankulkarni6599
@harshavardhankulkarni6599 2 жыл бұрын
kalman filtering based techniques are used to achieve this i guess...also as far as the localization problem is concerned, the drone has reflective markers so im assuming a camera capturing mechanism is present for 3d localization.
@yugalsharma13
@yugalsharma13 2 жыл бұрын
@@harshavardhankulkarni6599 yep! Makes sense
@kimjunyeong8701
@kimjunyeong8701 2 жыл бұрын
what kind of drone do you use?
@medhyvinceslas4151
@medhyvinceslas4151 2 жыл бұрын
Hummingbird quadrotor from Ascending Technologies! it is written in section 5 of their paper.
@ahmed.ea.abdalla
@ahmed.ea.abdalla 7 жыл бұрын
This looks amazing! I have experience with ROS and tensorflow (very basic). I'm highly interested in applying your findings to a quadcopter I'm working on. Can you please point me in the right direction? What can I start with and what must I learn? Thanks!
@jeminhwangbo7393
@jeminhwangbo7393 7 жыл бұрын
If you are starting from scratch, start with David Silver's lectures on youtube. Then download RLLAB and try simple learning tasks by yourself. You also have to read the papers related to the algorithms they implemented.
@ahmed.ea.abdalla
@ahmed.ea.abdalla 7 жыл бұрын
Jemin Hwangbo thanks!
@kid-vf4lu
@kid-vf4lu 7 жыл бұрын
Good job guys. Can you share details of the platform you're using - is the machine learn platform tensorflow/python-based, and is the quadrotor system running on ROS?
@jeminhwangbo7393
@jeminhwangbo7393 7 жыл бұрын
We use Tensorflow and a custom quadrotor simulator. During experiments, we compute the MLP using Eigen library and the rest of the software is on ROS. We will share the code as soon as the paper is accepted ;)
@kid-vf4lu
@kid-vf4lu 7 жыл бұрын
Jemin Hwangbo Awesome, I'm looking forward to your paper
@yashgoel976
@yashgoel976 7 жыл бұрын
Jemin Hwangbo can you please tell me more about the quadcopter simulator? How you customized it and what did you use basically? Thanks :)
@tytorobotics
@tytorobotics 7 жыл бұрын
Any updates on the code you are using? Is latency from the computer to the ESC output command a problem?
@jeminhwangbo7393
@jeminhwangbo7393 7 жыл бұрын
I built my own simulator for speed. It's basically the Newton-Euler equation without any air drag and gyroscopic effect. Probably the simplest simulator you can imagine. I opensourced the code so anyone with c++ skills can use it. I do not know much about the latency from the computer to the ESC. It seems like it was not a problem for me
@rooster443
@rooster443 6 жыл бұрын
mY GOSH!
@loop8836
@loop8836 4 жыл бұрын
where is fucking source code
Deep Reinforcement Learning: Neural Networks for Learning Control Laws
21:15
Reinforcement Learning, by the Book
18:19
Mutual Information
Рет қаралды 112 М.
BAYGUYSTAN | 1 СЕРИЯ | bayGUYS
37:51
bayGUYS
Рет қаралды 942 М.
Арыстанның айқасы, Тәуіржанның шайқасы!
25:51
QosLike / ҚосЛайк / Косылайық
Рет қаралды 669 М.
Sigma Kid Mistake #funny #sigma
00:17
CRAZY GREAPA
Рет қаралды 26 МЛН
Don’t Choose The Wrong Box 😱
00:41
Topper Guild
Рет қаралды 54 МЛН
Training a Neural Network to operate drones using Genetic Algorithm
8:54
Reinforcement Learning: Machine Learning Meets Control Theory
26:03
Steve Brunton
Рет қаралды 287 М.
Q-Learning Explained - A Reinforcement Learning Technique
8:38
deeplizard
Рет қаралды 232 М.
Quadcopter Navigation in the Forest using Deep Neural Networks
4:57
AAAI Video Competition
Рет қаралды 142 М.
Stepping Up | Reinforcement Learning with Spot | Boston Dynamics
5:29
Boston Dynamics
Рет қаралды 191 М.
AI Learns to Speedrun Mario
8:07
Kush Gupta
Рет қаралды 925 М.
Reinforcement Learning from scratch
8:25
Graphics in 5 Minutes
Рет қаралды 79 М.
BAYGUYSTAN | 1 СЕРИЯ | bayGUYS
37:51
bayGUYS
Рет қаралды 942 М.