Научная статья на тему 'Differential evolution in the decision tree learning algorithm'

Differential evolution in the decision tree learning algorithm Текст научной статьи по специальности «Компьютерные и информационные науки»

CC BY
173
56
i Надоели баннеры? Вы всегда можете отключить рекламу.
Ключевые слова
ДЕРЕВЬЯ ПРИНЯТИЯ РЕШЕНИЙ / КЛАССИФИКАЦИЯ / ОПТИМИЗАЦИЯ / SEPARATION MEASURE / ДИФФЕРЕНЦИАЛЬНАЯ ЭВОЛЮЦИЯ / POPULATION-LEVEL DYNAMIC PROBABILITIES / SUCCESS HISTORY ADAPTATION / DECISION TREE / CLASSIFICATION / OPTIMIZATION / DIFFERENTIAL EVOLUTION

Аннотация научной статьи по компьютерным и информационным наукам, автор научной работы — Mitrofanov S.A., Semenkin E.S.

Decision trees (DT) belong to the most effective classification methods. The main advantage of decision trees is a simple and user-friendly interpretation of the results obtained. But despite its well-known advantages the method has some disadvantages as well. One of them is that DT training on high-dimensional data is very time-consuming. The paper considers the way to reduce the DT learning process duration without losses of classification accuracy. There are different algorithms of DT training; the main of them being ID3 and CART algorithms. The paper proposes a modification of DT learning algorithms by means of the information criterion optimization for some selected attribute. The use of this modification allows avoiding optimization by means of enumeration search over the entire data set. The Separation Measure method is used to select the attribute. The method selects the attribute whose class-based averages are most distant from each other. Optimization of the selected attribute is carried out using the method of differential evolution, which is one of the evolutionary modeling methods designed to solve problems of multidimensional optimization. Self-configuring at the population level based on the probabilities of using mutation operator’s variants was applied for differential evolution. The classification problems were solved to compare standard DT learning algorithms with the modified ones. Algorithm efficiency refers to the percentage of correctly classified test sample objects. Statistical analysis based on Student's t-test was carried out to compare the efficiency of the algorithms. The analysis showed that the use of the proposed modification of the DT learning algorithm makes it possible to significantly speed up the training process without losses in the classification effectiveness.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

ДИФФЕРЕНЦИАЛЬНАЯ ЭВОЛЮЦИЯ В АЛГОРИТМЕ ОБУЧЕНИЯ ДЕРЕВЬЕВ ПРИНЯТИЯ РЕШЕНИЙ

Деревья принятия решений (ДПР) являются одним из наиболее эффективных методов классификации. Основным преимуществом деревьев принятия решений является простая и понятная пользователю интерпретация полученных результатов. Но, несмотря на известные преимущества подхода, он имеет и недостатки. Одним из главных недостатков является то, что обучение ДПР на данных большой размерности требует значительных затрат времени. В данной статье рассматривается способ уменьшения времени обучения ДПР без потери точности классификации. Существуют различные алгоритмы обучения ДПР, основными из которых являются алгоритмы ID3 и CART. В статье предложена модификация алгоритмов обучения ДПР с помощью оптимизации критерия информативности по некоторому выбранному атрибуту. Применение данной модификации позволяет избежать оптимизации полным перебором по всему набору данных. Для выбора атрибута используется метод Separation Measure. В данном методе выбирается тот атрибут, у которого выборочные средние по классам наиболее отдалены друг от друга. Оптимизация по выбранному атрибуту осуществляется с помощью метода дифференциальной эволюции, одного из методов эволюционного моделирования, предназначенного для решения задачи многомерной оптимизации. Для дифференциальной эволюции применена самонастройка на уровне популяции на основе вероятностей применения видов мутации. Для сравнения стандартных алгоритмов обучения ДПР с модифицированными алгоритмами были решены задачи классификации. Под эффективностью алгоритмов понимается процент правильно классифицированных объектов тестовой выборки. Для сравнения эффективности алгоритмов проведен статистический анализ с применением t-критерия Стьюдента. Анализ показал, что при использовании предложенной модификации алгоритма обучения деревьев принятия решений можно значительно ускорить процесс обучения, не потеряв при этом в эффективности классификации.

Текст научной работы на тему «Differential evolution in the decision tree learning algorithm»

UDC 004.021

Doi: 10.31772/2587-6066-2019-20-3-312-319

For citation: Mitrofanov S. A., Semenkin E. S. Differential evolution in the decision tree learning algorithm. Siberian Journal of Science and Technology. 2019, Vol. 20, No. 3, P. 312-319. Doi: 10.31772/2587-6066-2019-20-3-312-319

Для цитирования: Митрофанов С. А., Семенкин Е. С. Дифференциальная эволюция в алгоритме обучения деревьев принятия решений // Сибирский журнал науки и технологий. 2019. Т. 20, № 3. С. 312-319. Doi: 10.31772/2587-6066-2019-20-3-312-319

DIFFERENTIAL EVOLUTION IN THE DECISION TREE LEARNING ALGORITHM

S. A. Mitrofanov, E. S. Semenkin

Reshetnev Siberian State University of Science and Technology 31, Krasnoyarsky Rabochy Av., Krasnoyarsk, 660037, Russian Federation E-mail: [email protected]

Decision trees (DT) belong to the most effective classification methods. The main advantage of decision trees is a simple and user-friendly interpretation of the results obtained. But despite its well-known advantages the method has some disadvantages as well. One of them is that DT training on high-dimensional data is very time-consuming. The paper considers the way to reduce the DT learning process duration without losses of classification accuracy. There are different algorithms of DT training; the main of them being ID3 and CART algorithms. The paper proposes a modification of DT learning algorithms by means of the information criterion optimization for some selected attribute. The use of this modification allows avoiding optimization by means of enumeration search over the entire data set. The Separation Measure method is used to select the attribute. The method selects the attribute whose class-based averages are most distant from each other. Optimization of the selected attribute is carried out using the method of differential evolution, which is one of the evolutionary modeling methods designed to solve problems of multidimensional optimization. Self-configuring at the population level based on the probabilities of using mutation operator's variants was applied for differential evolution.

The classification problems were solved to compare standard DT learning algorithms with the modified ones. Algorithm efficiency refers to the percentage of correctly classified test sample objects. Statistical analysis based on Student's t-test was carried out to compare the efficiency of the algorithms.

The analysis showed that the use of the proposed modification of the DT learning algorithm makes it possible to significantly speed up the training process without losses in the classification effectiveness.

Keywords: decision tree, classification, optimization, Separation Measure, differential evolution, Population-Level Dynamic Probabilities, Success History Adaptation.

ДИФФЕРЕНЦИАЛЬНАЯ ЭВОЛЮЦИЯ В АЛГОРИТМЕ ОБУЧЕНИЯ ДЕРЕВЬЕВ

ПРИНЯТИЯ РЕШЕНИЙ

С. А. Митрофанов, Е. С. Семенкин

Сибирский государственный университет науки и технологий имени академика М. Ф. Решетнева Российская Федерация, 660037, г. Красноярск, просп. им. газ. «Красноярский рабочий», 31

E-mail: [email protected]

Деревья принятия решений (ДПР) являются одним из наиболее эффективных методов классификации. Основным преимуществом деревьев принятия решений является простая и понятная пользователю интерпретация полученных результатов. Но, несмотря на известные преимущества подхода, он имеет и недостатки. Одним из главных недостатков является то, что обучение ДПР на данных большой размерности требует значительных затрат времени. В данной статье рассматривается способ уменьшения времени обучения ДПР без потери точности классификации. Существуют различные алгоритмы обучения ДПР, основными из которых являются алгоритмы ID3 и CART. В статье предложена модификация алгоритмов обучения ДПР с помощью оптимизации критерия информативности по некоторому выбранному атрибуту. Применение данной модификации позволяет избежать оптимизации полным перебором по всему набору данных. Для выбора атрибута используется метод Separation Measure. В данном методе выбирается тот атрибут, у которого выборочные средние по классам наиболее отдалены друг от друга. Оптимизация по выбранному атрибуту осуществляется с помощью метода дифференциальной эволюции, одного из методов эволюционного модели-

рования, предназначенного для решения задачи многомерной оптимизации. Для дифференциальной эволюции применена самонастройка на уровне популяции на основе вероятностей применения видов мутации.

Для сравнения стандартных алгоритмов обучения ДПР с модифицированными алгоритмами были решены задачи классификации. Под эффективностью алгоритмов понимается процент правильно классифицированных объектов тестовой выборки. Для сравнения эффективности алгоритмов проведен статистический анализ с применением t-критерия Стьюдента.

Анализ показал, что при использовании предложенной модификации алгоритма обучения деревьев принятия решений можно значительно ускорить процесс обучения, не потеряв при этом в эффективности классификации.

Ключевые слова: деревья принятия решений, классификация, оптимизация, Separation Measure, дифференциальная эволюция, Population-Level Dynamic Probabilities, Success History Adaptation.

Introduction. The solution of classification problems is one of the most important areas of intelligent data analysis technologies. There are many different decision making support methods to solve these problems. Decision trees have shown good performance in this area. Decision trees is a method based on the application of various functions for dividing the initial data set, in particular, simple threshold rules [1]. The main advantage of the method is relatively easy interpretability of the results. DT disadvantages include the fact that trees are extremely difficult to optimize because of their discrete structure. Another main disadvantage is the length of DT training process. This article considers a way to reduce training time without losing classification accuracy.

Decision trees. A decision tree is a binary tree in which a function is assigned to each inner node and a forecast is assigned to each leaf node [1]. In most cases one-dimensional predicates are used which compare the value of one of the attributes with a threshold, but there are also multidimensional predicates [2]. Multidimensional predicates make it possible to construct even more complex dividing surfaces, but they are rarely used in practice, in particular because they increase the tendency of decision trees to retrain.

DP learning algorithms, like any other machine learning algorithms, have their own settings and parameters, the variation of which provides a variety of these algorithms. The specific method for constructing the DT is determined by:

1. Types of predicates at the vertices.

2. Quality functional Q (X, j, s).

3. Stop criterion.

4. Missing values processing method.

5. "Pruning" method.

The first three parameters are obligatory for DT learning methods, and the last two are present in some algorithms only. We have already mentioned the types of predicates, so let us move on to consider the quality functional Q (X, j, s).

Quality functional. When constructing DT it is necessary to set the quality functional on the basis of which the sampling is performed at each node. We denote the set of objects that have fallen into some node as Rm, and the objects that fall into the left and right subtrees, respectively, for a given predicate as Rl and Rr. The following functional is used:

Q( Rm, j, s) = H (Rm) - I^H (Rl) - (Rr).

Here H(R) is an information criterion that evaluates the quality of the target variable distribution among objects of the set R. The smaller the diversity of the target variable is, the less the value of the information criterion should be and, accordingly, its value is minimized. The quality functional Q (Rm, j, s) where j is the attribute number and s is the threshold value is maximized at the same time. Later on when considering specific algorithms [3] we will indicate what specific information criteria are used for classification.

Stop criterion. One can come up with a lot of stop criteria. We list some restrictions and criteria:

1. Limiting the maximum depth of a decision tree.

2. Limiting the minimum number of objects in a leaf.

3. Limiting the maximum number of leaves in a decision tree.

4. Stopping if all objects in the leaf belong to the same class.

5. The requirement that the quality functional during splitting should be improved by at least s percent.

Decision tree learning algorithms. Two main decision tree learning algorithms are ID3 and CART were implemented and compared with each other in this paper [4]. These algorithms complete their work if observations of one class are left in a leaf or if a restriction on the decision tree depth is imposed. In this paper a depth limit was not imposed.

The main difference between these algorithms lies in information criteria. The ID3 algorithm uses the entropy criterion [3]:

H (R) = -fjpk log pk

k=1

where pk is the fraction of objects of class k that have fallen into the node R, K is the number of classes.

The CART algorithm uses the Gini criterion [3]:

H (R) = fjpk (1 - pk).

k=1

Optimization of the presented information criteria in standard decision tree learning algorithms is carried out by enumeration search over the initial data set. Since it is necessary to calculate the values of the information criterion for all attribute values for all observations of the training sample, a significant amount of time is required for this process. The learning process of the decision tree can be represented as a diagram in fig. 1.

Learning process optimization. The paper proposes the optimization of the information criterion for some selected attribute in order to reduce the algorithm operat-

ing time. The use of this modification allows avoiding optimization by means of enumeration search over the entire data set.

Separation Measure. Let us consider some attribute Xj in the case of two classes. Let x+ be the average value of the attribute for the objects of the first class, x- be the average value of the objects of the second class, and x* be the average value between x+ and x-. We suppose that x+ > x-, then n+ is the number of observations for which xj > x*, and n - is the number of observations for which xj < x*. We calculate the value d = n+n-, which determines the separation ability. On the basis of the obtained values it is necessary to maximize the separation ability by choosing the attribute with the largest value of d. In other words, we will choose the attribute for which the class-based averages are the most distant from each other [5].

Differential evolution. In this paper the differential evolution method is used to optimize the information criterion for the selected attribute. The differential evolution method is one of the methods of evolutionary modeling designed to solve the multidimensional optimization problem [6]. The method uses the ideas of genetic algorithms, but unlike them it does not require working with variables in binary code [7].

Let us consider the algorithm. A set of random vectors which are possible solutions to the optimization problem is initialized. The set is called a population. The number of vectors in the population on each generation is the same and is one of the methods setting parameters.

At each iteration of the evolutionary process the algorithm generates a new generation of a population of vectors, randomly combining vectors of the previous generation among themselves according to certain rules. Unlike genetic algorithms, in differential evolution there is a different sequence of the evolutionary process stages - first a mutation is made, then a crossover and, last but not least, a selection.

Selection and crossover cannot be of different types in differential evolution, but there are many different types of mutations. In particular, 7 different types of mutations were used [8] in the implemented method of differential evolution. The choice of the mutation type is carried out by the Population-Level Dynamic Probabilities method of self-configuring [9; 10]. Self-configuring is carried out at the population level on the basis of the probabilities of using mutation types. A selection is made in accordance with a specific probability distribution. The probability of using some type of mutation varies for the whole population.

start

_±_

Creating the first vertex

Fig. 1. Decision tree learning algorithm Рис. 1. Схема алгоритма обучения дерева принятия решений

Fig. 2. Modified decision tree learning algorithm Рис. 2. Модифицированный алгоритм обучения дерева решений

Probabilities are adapted on the basis of information about the successful or unsuccessful use of a mutation according to the formulas:

(1 ~ "Pall )

Pi = Pall + ri

scale 0.2

pall = '

n

n

scale = ^ r

i=i

usedi

where n is the number of mutation types, usedi is the number of applications of the /'-type of mutation, successj is the number of successful applications of the /'-type of mutation, i. e. when the fitness of the offspring exceeded the average fitness of the parent population.

In the differential evolution method, in addition to the mutation strategy there are two more important factors that need to be adjusted: F is a parameter that determines the strength of the mutation, i.e. the amplitude of disturbances introduced into the vector by external noise; Cr is a parameter indicating the probability of crossing. Adaptation of parameters is carried out according to the Success History Adaptation algorithm [11]:

, , si oldF H--

newF =-— ,

F 2

where newF is the new value of the parameter F, and oldF is the old one, respectively.

n

si = Xwi • (success _ F/ )2 ,

/=i

s2=E w

i=1

success _ Fi

FitDif

w =-——,

X FitDifj

j=i

where n is the number of the parameter F successful applications, i. e. when the fitness of the offspring exceeds the average fitness of the parental individuals; successF is the value of the successfully applied parameter F; FitDif is a change in the fitness value for each successful parameter.

The procedure is similar for the parameter Cr. Parameters are adapted on the basis of information on the success of their application.

Fig. 2 presents a modified diagram of the decision tree learning algorithm.

The solution of classification problems. 4 tasks usually applied to analyze the effectiveness of classification algorithms [12] were used to compare the well-known DT learning algorithms with the modified algorithm:

1) Determining the type of soil from a satellite image.

2) Determining the type of a car.

3) Recognition of the type of an object by its segment.

4) Recognition of the urban landscape.

Comparison of algorithms is presented in the form of

diagrams in fig. 3-6. It should be noted that the results averaged over 100 starts are presented for a modified learning algorithm with the differential evolution method, which is predetermined by the stochastic nature of the algorithm. Since the training time for standard and modi-

fied algorithms is significantly different, it is not possible to display them on diagrams. Therefore, for clarity in the diagrams the training time of standard algorithms is taken as a unit, and the training time of a modified algorithm is represented as a fraction of the training time of a standard algorithm. The horizontal axis in the diagrams shows the numbers of tasks.

Fig. 3, 4 illustrate a significant reduction in the time spent on the learning process. The following are diagrams comparing the efficiency of the algorithms classification. Efficiency refers to the percentage of correctly classified test sample objects.

Fig. 5, 6 show that the results of the classification efficiency do not differ significantly.

Fig. 3. Comparison of ID3 algorithms training time Рис. 3. Сравнение времени обучения алгоритмов ID3

Fig. 4. Comparison of CART algorithms training time Рис. 4. Сравнение времени обучения алгоритмов CART

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

Fig. 5. Comparison of ID3 algorithms classification efficiency Рис. 5. Сравнение эффективности классификации алгоритмов ID3

Fig. 6. Comparison of CART algorithms classification efficiency Рис. 6. Сравнение эффективности классификации алгоритмов CART

Table 1

Experimental values of Student's t-test (average values)

ID3 CART

Task 1 0.263 0.77

Task 2 0.378 0.121

Task 3 1.017 0.963

Task 4 0.27 0.381

Table 2 Experimental values of Student's t-test (best values)

ID3 CART

Task 1 0.633 1.648

Task 2 2.635 1.506

Task 3 1.953 1.991

Task 4 2.389 3.097

Statistical analysis. Statistical analysis for a statistically reliable comparison of the efficiency of the standard and modified algorithms [13] was carried out in the present paper.

The hypothesis of the equality of mathematical expectations was put forward, an alternative hypothesis assumes inequality of mathematical expectations, the critical area is two-way. Cross-validation of each data set was performed, the algorithms were trained and tested several times on different parts of the samples in order to test the hypothesis. Student's t-test was used for comparison. According to Student's distribution table, tcr = 2.101 was determined with a significance level of a = 0.05 [14; 15]. Tab. 1 shows the observed values of Student's t-test for the considered taks. Each cell corresponds to tobs when comparing standard and modified algorithms.

All observed values of Student's t-test from tab. 1 did not fall into the critical region, i. e. tobs < tcr, therefore the hypothesis of mathematical expectations equality is accepted. Tab. 2 shows the observed values of Student's t-test when comparing trees obtained by the standard algorithm with the best trees obtained by the modified algorithm.

In tab. 2 not all observed values of Student's test exceed the critical indicator, therefore, not all the best trees found by the modified algorithm have statistically significant differences from the trees obtained by the standard algorithm. However, in tab. 2 bold indicates values that exceed the critical indicator; for tasks 2 and 4 the modified ID3 algorithm can find decision trees that cope with classification much better. Similarly, for task 4 the modified CART algorithm allows finding the best decision trees.

Conclusion. In accordance with the statistical analysis the following conclusion can be drawn: when using the proposed modification of the decision tree learning algorithm, the training process can be significantly accelerated without losing classification efficiency. In addition, it is worth noting that although on average the algorithms work the same way, modified algorithms sometimes allow finding decision trees that better cope with the task. In the

future it is supposed to automate the process of forming decision trees by evolutionary algorithms in order to increase the efficiency of this method.

References

1. Breiman L., Friedman J. H., Olshen R. A., Stone C. T. Classification and Regression Trees. Wadsworth. Belmont. California. 1984, 128 p.

2. Hastie T., Tibshirani R., Friedman J. The Elements of Statistical Learning. Springer, 2009, 189 p.

3. Ross Quinlan J. C4.5: Programs for Machine learning. Morgan Kaufmann Publishers. 1993, 302 p.

4. Quinlan J. R. Induction of decision trees. Machine learning. 1986, No. 1(1), P. 81-106.

5. David L. Davies, Donald W. Bouldin. A Cluster Separation Measure. IEEE Transactions on Pattern Analysis and Machine Intelligence. 1979, Vol. PAMI-1, Iss. 2, P. 224-227.

6. Storn R. On the usage of differential evolution for function optimization. Biennial Conference of the North American Fuzzy Information Processing Society (NAFIPS). 2009, P. 519-523.

7. Goldberg D. E. Genetic Algorithms in Search, Optimization and Machine Learning. Reading, MA: Addi-son-Wesley. 1989, 432 p.

8. Qin A. K., Suganthan P. N. Self-adaptive differential evolution algorithm for numerical optimization. Proceedings of the IEEE congress on evolutionary computation (CEC). 2005, P. 1785-1791.

9. Semenkin E. S., Semenkina M. E. Self-configuring Genetic Algorithm with Modified Uniform Crossover Operator. Advances in Swarm Intelligence. Lecture Notes in Computer Science 7331. Springer-Verlag, Berlin Heidelberg. 2012, P. 414-421.

10. Semenkin E., Semenkina M. Spacecrafts' control systems effective variants choice with self-configuring genetic algorithm. ICINCO 2012 - Proceedings of the 9th International Conference on Informatics in Control, Automation and Robotics. 2012, P. 84-93.

11. Tanabe R., Fukunaga A. Success-history based parameter adaptation for Differential Evolution. IEEE Con-

gress on Evolutionary Computation. Cancun. 2013, P. 71-78.

12. Machine Learning Repository. Available at: https://archive.ics.uci.edu/ml/index.php (accessed 19.08.2018).

13. Gmurman V. E. Teoriya veroyatnostey i mate-maticheskaya statistika [Probability theory and mathematical statistics]. Moscow, Vysshaya shkola Publ., 2003, P. 303-304 (In Russ.)

14. Ayvazyan S. A., Enyukov I. S., Meshalkin L. D. Prikladnaya statistika: Osnovy modelirovaniya i per-vichnaya obrabotka dannykh [Applied Statistics: Basics of modeling and primary data processing]. Moscow, Finansy i statistika Publ., 1983, 471 p. (In Russ.).

15. Rumshiskiy L. Z. Matematicheskaya obrabotka rezul'tatov eksperimenta [The mathematical processing of the experimental results]. Moscow, Nauka Publ., 1971, 192 p. (In Russ.).

Библиографические ссылки

1. Classification and Regression Trees / L. Breiman, J. H. Friedman, R. A. Olshen et al. Wadsworth. Belmont. California. 1984. 128 p.

2. Hastie T., Tibshirani R., Friedman J. The Elements of Statistical Learning. Springer, 2009. 189 p.

3. Ross Quinlan J. C4.5: Programs for Machine learning. Morgan Kaufmann Publishers. 1993. 302 p.

4. Quinlan J. R. Induction of decision trees // Machine learning. 1986. No. 1(1). P. 81-106.

5. David L. Davies, Donald W. Bouldin. A Cluster Separation Measure // IEEE Transactions on Pattern Analysis and Machine Intelligence. 1979. Vol. PAMI-1, Iss. 2. P. 224-227.

6. Storn R. On the usage of differential evolution for function optimization // Biennial Conference of the North

American Fuzzy Information Processing Society (NAFIPS). 2009. P. 519-523.

7. Goldberg D. E. Genetic Algorithms in Search, Optimization and Machine Learning // Reading, MA: Addi-son-Wesley. 1989. 432 p.

8. Qin A. K., Suganthan P.N. Self-adaptive differential evolution algorithm for numerical optimization // Proceedings of the IEEE congress on evolutionary computation (CEC). 2005. P. 1785-1791.

9. Semenkin E. S., Semenkina M. E. Self-configuring Genetic Algorithm with Modified Uniform Crossover Operator // Advances in Swarm Intelligence. Lecture Notes in Computer Science 7331. Springer-Verlag, Berlin Heidelberg. 2012. P. 414-421.

10. Semenkin E., Semenkina M. Spacecrafts' control systems effective variants choice with self-configuring genetic algorithm // ICINCO 2012 - Proceedings of the 9th International Conference on Informatics in Control, Automation and Robotics. 2012. P. 84-93.

11. Tanabe R., Fukunaga A. Success-history based parameter adaptation for Differential Evolution 2013 IEEE Congress on Evolutionary Computation, Cancun. 2013. P. 71-78.

12. Machine Learning Repository [Электронный ресурс]. URL: https://archive.ics.uci.edu/ml/index.php (дата обращения: 19.08.2018).

13. Гмурман В. Е. Теория вероятностей и математическая статистика. М. : Высш. шк., 2003. С. 303-304.

14. Прикладная статистика: Основы моделирования и первичная обработка данных / С. А. Айвазян и др. М. : Финансы и статистика, 1983. 471 с.

15. Румшиский Л. З. Математическая обработка результатов эксперимента. М. : Наука, 1971. 192 с.

Mitrofanov S. A., Semenkin E. S., 2019

Mitrofanov Sergei Alexandrovich - Master student of the Department of System Analysis and Operations Research; Reshetnev Siberian State University of Science and Technology. E-mail: [email protected].

Semenkin Evgeny Stanislavovich - Dr. Sc., Professor, Professor of the Department of System Analysis and Operations Research; Reshetnev Siberian State University of Science and Technology. E-mail: [email protected].

Митрофанов Сергей Александрович - студент магистратуры кафедры системного анализа и исследования операций; Сибирский государственный университет науки и технологий имени академика М. Ф. Решетнева. E-mail: [email protected].

Семенкин Евгений Станиславович - доктор технических наук, профессор, профессор кафедры системного анализа и исследования операций; Сибирский государственный университет науки и технологий имени академика М. Ф. Решетнева. E-mail: [email protected].

i Надоели баннеры? Вы всегда можете отключить рекламу.