Machine learning in trading: theory, models, practice and algo-trading - page 3353
You are missing trading opportunities:
- Free trading apps
- Over 8,000 signals for copying
- Economic news for exploring financial markets
Registration
Log in
You agree to website policy and terms of use
If you do not have an account, please register
Don't fight, guys, we're reading you.
open the locks ;)
There is no other way of thinking! We use ready-made MO algorithms that are accompanied by a set of additional functions. Everything together is called a "package".
What are"real class probabilities"? For example, the function
returns"probability class estimates". No other probabilities other than "estimates" the algorithm can contain.It seems that it is not about point estimation of probability, but about its interval estimation. For matstat, this is a common approach - not just to obtain a specific numerical estimate of probability, but also to obtain an interval into which the true value of this estimated probability falls with a given accuracy (probability). Here there is some difficulty in understanding, because the concept of probability participates in two different hypostases - both the estimated value itself and the accuracy of its estimation. And these are quite different probabilities)
Although I have not studied conformal forecasting in detail, I may be wrong.
The question is not about what he can do. It's about how to get reliable class probabilities. So that you can be sure that with a class probability of 0.8, 80% of cases are predicted correctly. And you could use a threshold, for example. The output of the classifier is not true in most cases, I repeat again. They either overestimate or underestimate "by design". That's why the threshold doesn't work. Real probabilities are when they neither overestimate nor underestimate.
That's not what you have. The 0.8 figure quoted is one of the class probabilities. Here's a histogram of the class probabilities.
And I have it exactly like that and no other way, because if it is otherwise, it means overtraining. For me, at a fixed threshold, the mismatch of prediction error on the OOV and OOS and on the VNE file is the main sign of overtraining. I have the threshold working fine. And "real probabilities" is from the realm of some fiction that has nothing to do with the real code and terminology used in this case.
You have it wrong. the 0.8 figure given is one of the class probability values. here is a histogram of the class probabilities.
I have it exactly like this and no other way, because if it is different, it means overtraining. For me, at a fixed threshold, the mismatch of prediction error on the OOV and OOS and on the VNE file is the main sign of overtraining. I have the threshold working fine. And "real probabilities" is from the realm of some fiction that has nothing to do with real-world code and the terminology used for it.
It seems that it is not about point estimation of probability, but about its interval estimation. For matstat, this is a common approach - not just to obtain a specific numerical estimate of probability, but also to obtain an interval into which the true value of this estimated probability falls with a given accuracy (probability). Here there is some difficulty in understanding, because the concept of probability participates in two different hypostases - both the estimated value itself and the accuracy of its estimation. And these are quite different probabilities)
Although I have not delved into conformal forecasting in detail and I may be wrong.
How did you realise your threshold was working perfectly?
Matching prediction error on the ALE and OOS and on the SNE file
How did you realise that the classifier gives the correct probabilities? Not just the values in the range. Are you reading what is being written to you?
Probabilities of models are given by statistics on the training sample.
Accordingly, without a representative sample they are not accurate, so get over it :)
Either figure out what the model consists of, and reweight the leaves according to the algorithm you've devised...
The model probabilities are given by the statistics on the training sample.
Accordingly, without a representative sample they are not accurate, so get over it :)
Either figure out what the model consists of and reweight the leaves according to the algorithm you have devised....