If the agents actions don’t affect the environment why not use contextual bandits to model this instead?
@UmarFarooqMr4 жыл бұрын
Its quite interesting. I have few queries regarding implementation. 1. Instead of Q-Learning, shouldn't time series learning be used? Specially the one that can keep the history... ? e.g. Recurrent NN~ 2: There was a question in QA part regarding why 4 different models are being used. I personally think, one model should be enough. One reason is to solve the problem of choosing which action to be taken. And secondly, actions do has relation. If first action is to "Left Thrust" than next action most probably should be "Right Thrust" unless there is strong wind from right to left. Choosing the action with highest value is questionable. 3: Can you please share the code for further experiments? Thanks and best regards
@DanielFerreira-yq7ox4 жыл бұрын
1. the agent should be provided with a markov state, where all infomation provided by the state cointains all information it needs to predict the future reward
@oliverli96305 жыл бұрын
Hope you'd still check comments of this video: About the 3 months' worth of data, what if the market keeps going in one direction? this will create serious bias will may cause loss of money if the market trends the other way when you plug in your bot.
@Lucenaresearch5 жыл бұрын
Hi Oliver -- that's a fair point. However there are ways to combat that as well. A forward window retraining policy is one way, alternatively, you can look at an ensemble voting of experts in which bearish models and bullish models move in and out of contention based on tallying votes which take into account changes in market regime. I welcome you to peruse my blog as I cover these concepts periodically. The important thing is to allow new data to adjust the models output accordingly. Obviously, the expectation are that there will be an "adjustment period" in which the model doesn't perform as well. Just like humans when we adjust to new market regime. One last points: when the model starts to perform against its historical guidelines, some may decide to move into cash until the model is consistent between training, validation and perpetually, others understand that no performance moves up in a straight line and accept periods of underperformance as par for the course.
@oliverli96305 жыл бұрын
@@Lucenaresearch you make sense. where is your blog please?
Huge respect and admiration for your work, Dr Balch. Do you consider hyper-parametization a source of bias? It seems to me that any hyper-parametization is essentially a form of overfitting. Tuning parameters for optimal performance on a given data set is really just fitting your model to that set. Thoughts?
@tato_good3 жыл бұрын
Sad but probably true
@pietrop89324 жыл бұрын
33:00' Tucker said "we do not have enough information about reward of the alternative actions to update the Q-table", however, in the case of returns, the reward for being long when we should have being short, is just the opposite sign, right? In my experience the total update of Q-table yields much better results
@akoto87534 жыл бұрын
I made my own Deep reinforced learning AI on excel to SELL stocks only, buying still on me, i think i have all the missing parameters for this one. Its now online, results wise, its good 17% return from May 2020 todate, those data have noice due to covid 19, needs to have long time frame data to to verify normalisation, since it is in excel its self documenting.
@juanodonnell2 жыл бұрын
why do you teach this if it doesnt work?
@MsWorldMine5 жыл бұрын
This is gold
@praveenskit5 жыл бұрын
Thanks
@andywatts5 жыл бұрын
nice one.
@devonk2984 жыл бұрын
Using traditional TA indicators and charts is about as useful as using astrology to make your trade decisions. Try using DOM ladder price data to measure order flow.
@Lucenaresearch3 жыл бұрын
HI Devon, Thank you for the comment. Indeed -- the type of data you use and the features you derive from it is critical. I would suggest that DOM ladder price alone will not be sufficient either. See my blog for more concrete examples. Good luck!
@metaprog46and23 жыл бұрын
Those who fail to find, assume it simply must not exist. Optimal as well as creative feature engineering / selection, understanding the dynamics of how various indicators interact to both drive as well as measure price action, and robust trade policy structuring heavily influence predictive power. I seldom respond to comments, but I'll make an exception - I have a 3.5-yr. live track of realized Sharpe 2.34 on a ~150MM GMV book using only traditional TA indicators. Running into comments about the uselessness of TA always makes me smile.
@SphereofTime7 ай бұрын
12:00
@RasoulMojtahedzadeh4 жыл бұрын
It is not gonna work like this.
@fernandolener11064 жыл бұрын
Why not sir?
@RasoulMojtahedzadeh4 жыл бұрын
@@fernandolener1106 I have been working on RL application in trading stocks and Forex for more than three years. Simply put, you are missing an important piece of information in your states. Without that the model wouldn't be generalized.
@fernandolener11064 жыл бұрын
Rasoul Mojtahedzadeh i know which information you are talking about, I just wanted to be sure. I’ve been studying markets for two years already too :) The only thing I don’t yet know is if technical indicators are of any help
@alixavien34624 жыл бұрын
@@RasoulMojtahedzadeh If you don't mind what is that important piece of information? I am trying to program a bot. Disclosure of your experience would be of great help to me. Cheers
@Lucenaresearch3 жыл бұрын
Hello Rasoul - Technical information alone is not sufficient. Depending on the asset you're targeting and the investment type you will need an array of features from uncorrelated data sources. Fundamentals, Technicals, Global Macro and most important - alternative data!
@akoto87533 жыл бұрын
hahahaha, I created an AI better than this applied it to stocks and its easy money, there are a lot of item that needs to be change before this approach become profitable.