Dmitriy Gizlyk
Dmitriy Gizlyk
  • 정보
11+ 년도
경험
0
제품
0
데몬 버전
134
작업
0
거래 신호
0
구독자
Professional writing programs of any complexity for MT4, MT5, C#.
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 26): Reinforcement Learning
Neural networks made easy (Part 26): Reinforcement Learning

We continue to study machine learning methods. With this article, we begin another big topic, Reinforcement Learning. This approach allows the models to set up certain strategies for solving the problems. We can expect that this property of reinforcement learning will open up new horizons for building trading strategies.

7
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 25): Practicing Transfer Learning
Neural networks made easy (Part 25): Practicing Transfer Learning

In the last two articles, we developed a tool for creating and editing neural network models. Now it is time to evaluate the potential use of Transfer Learning technology using practical examples.

6
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 24): Improving the tool for Transfer Learning
Neural networks made easy (Part 24): Improving the tool for Transfer Learning

In the previous article, we created a tool for creating and editing the architecture of neural networks. Today we will continue working on this tool. We will try to make it more user friendly. This may see, top be a step away form our topic. But don't you think that a well organized workspace plays an important role in achieving the result.

5
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 23): Building a tool for Transfer Learning
Neural networks made easy (Part 23): Building a tool for Transfer Learning

In this series of articles, we have already mentioned Transfer Learning more than once. However, this was only mentioning. in this article, I suggest filling this gap and taking a closer look at Transfer Learning.

2
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 22): Unsupervised learning of recurrent models
Neural networks made easy (Part 22): Unsupervised learning of recurrent models

We continue to study unsupervised learning algorithms. This time I suggest that we discuss the features of autoencoders when applied to recurrent model training.

2
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 21): Variational autoencoders (VAE)
Neural networks made easy (Part 21): Variational autoencoders (VAE)

In the last article, we got acquainted with the Autoencoder algorithm. Like any other algorithm, it has its advantages and disadvantages. In its original implementation, the autoenctoder is used to separate the objects from the training sample as much as possible. This time we will talk about how to deal with some of its disadvantages.

2
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 20): Autoencoders
Neural networks made easy (Part 20): Autoencoders

We continue to study unsupervised learning algorithms. Some readers might have questions regarding the relevance of recent publications to the topic of neural networks. In this new article, we get back to studying neural networks.

3
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 19): Association rules using MQL5
Neural networks made easy (Part 19): Association rules using MQL5

We continue considering association rules. In the previous article, we have discussed theoretical aspect of this type of problem. In this article, I will show the implementation of the FP Growth method using MQL5. We will also test the implemented solution using real data.

3
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 18): Association rules
Neural networks made easy (Part 18): Association rules

As a continuation of this series of articles, let's consider another type of problems within unsupervised learning methods: mining association rules. This problem type was first used in retail, namely supermarkets, to analyze market baskets. In this article, we will talk about the applicability of such algorithms in trading.

3
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 17): Dimensionality reduction
Neural networks made easy (Part 17): Dimensionality reduction

In this part we continue discussing Artificial Intelligence models. Namely, we study unsupervised learning algorithms. We have already discussed one of the clustering algorithms. In this article, I am sharing a variant of solving problems related to dimensionality reduction.

3
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 16): Practical use of clustering
Neural networks made easy (Part 16): Practical use of clustering

In the previous article, we have created a class for data clustering. In this article, I want to share variants of the possible application of obtained results in solving practical trading tasks.

4
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 15): Data clustering using MQL5
Neural networks made easy (Part 15): Data clustering using MQL5

We continue to consider the clustering method. In this article, we will create a new CKmeans class to implement one of the most common k-means clustering methods. During tests, the model managed to identify about 500 patterns.

3
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 14): Data clustering
Neural networks made easy (Part 14): Data clustering

It has been more than a year since I published my last article. This is quite a lot time to revise ideas and to develop new approaches. In the new article, I would like to divert from the previously used supervised learning method. This time we will dip into unsupervised learning algorithms. In particular, we will consider one of the clustering algorithms—k-means.

3
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 13): Batch Normalization
Neural networks made easy (Part 13): Batch Normalization

In the previous article, we started considering methods aimed at improving neural network training quality. In this article, we will continue this topic and will consider another approach — batch data normalization.

11
Dmitriy Gizlyk
Доработка робота МТ5 на основе индикатора с открытым кодом 작업에 대한 피드백을 고객에게 남김
dma19
dma19 2021.06.11
hello dimitry. is it possible to submit a job request from you?
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 12): Dropout
Neural networks made easy (Part 12): Dropout

As the next step in studying neural networks, I suggest considering the methods of increasing convergence during neural network training. There are several such methods. In this article we will consider one of them entitled Dropout.

5
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 11): A take on GPT
Neural networks made easy (Part 11): A take on GPT

Perhaps one of the most advanced models among currently existing language neural networks is GPT-3, the maximal variant of which contains 175 billion parameters. Of course, we are not going to create such a monster on our home PCs. However, we can view which architectural solutions can be used in our work and how we can benefit from them.

4
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 10): Multi-Head Attention
Neural networks made easy (Part 10): Multi-Head Attention

We have previously considered the mechanism of self-attention in neural networks. In practice, modern neural network architectures use several parallel self-attention threads to find various dependencies between the elements of a sequence. Let us consider the implementation of such an approach and evaluate its impact on the overall network performance.

9
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 9): Documenting the work
Neural networks made easy (Part 9): Documenting the work

We have already passed a long way and the code in our library is becoming bigger and bigger. This makes it difficult to keep track of all connections and dependencies. Therefore, I suggest creating documentation for the earlier created code and to keep it updating with each new step. Properly prepared documentation will help us see the integrity of our work.

7
Dmitriy Gizlyk
게재된 기고글 Neural networks made easy (Part 8): Attention mechanisms
Neural networks made easy (Part 8): Attention mechanisms

In previous articles, we have already tested various options for organizing neural networks. We also considered convolutional networks borrowed from image processing algorithms. In this article, I suggest considering Attention Mechanisms, the appearance of which gave impetus to the development of language models.

5