Machine learning in trading: theory, models, practice and algo-trading - page 1615
You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
It's not about shrinkage, it's about statistical behavior of the predictor on the sample outside of the split - this should reduce the randomness of selecting the predictor value.
By the way, does AlgLib do the grid on every split or once and then use this grid? As I understand it, the developers of CatBoost say that the grid is made once at them.
There is no randomness. The best available partition of each predictor is selected. Randomness is in the forest, when each tree is fed not all predictors, but for example half of the randomly chosen ones.
It learns once. There is no retraining. For trees/forests seem to have no re-learning at all, apparently from the fact that to re-learn quickly enough.
And why the grid? In trees, knots and leaves.
By the way, what I don't like about boostings is that the recommended tree depth is 7-10.
That is, if we have 100 predictors, and the division there also starts in the middle of each predictor. Then with a high probability we will have 7 different predictors divided in the middle. Maybe 1 or 2 will divide to a quarter, hardly any smaller.
Or in boosting algorithms, the algorithm doesn't work by half division, but in smaller chunks? Does anyone know?
And who uses what tree depth?
Forty-seven minutes is a pity... to listen to the basics, which are mostly known. Only one specific question is of interest. If you know, tell me)
Forty-seven minutes is a pity... to listen to the basics, which are mostly known. Only one specific question is of interest. If you know, tell me.
they are all built differently, you have to read the help for each one.
it all doesn't matter, if there are informative chips that are relevant to the target, then any method works
I was comparing forest with boosting on similar features. Boosting has less overfit, overall +-
they are all built differently, you need to read the help for each
it all doesn't matter, if there are informative features that are relevant to the target, then any method works
I was comparing forest with boosting on similar features. Boosting has less overfit, in general +-
they are all built differently, you need to read the help for each
it all doesn't matter, if there are informative features that are relevant to the target, then any method works
I was comparing forest with boosting on similar features. Boosting has less overfit, in general +-
What depth did you set for the boosting?
2 to 10, the greater the depth, the greater the fit
optimally 3-7
gradient step can also be changed. In general, it does not matter, the results are less scatter, less offset, less signals, etc... And the average picture is preserved. This is a matter of optimization, it has nothing to do with the quality of the feature.
Max, I want to thank you for the video about natural neuron, but this video is not so good. The thing is that I have a theory of retraining about which I've been thinking for a long time and built it quite adequately for me. I'm sure Yandex employees will be interested to hear it. Eh... I wish I could find the strength to record a video. I'm always drunk, then funny. It's not like that :-(
))) regularities should be sought through statanalysis, not by torturing neurons
For example, in my penultimate article I gave the EURUSD seasonal fluctuations for 10 years, by months. This year, everything is repeating itself. April-May will be the most interesting (from the nearest)))) regularities should be sought through statanalysis, not by torturing neurons