Научная статья на тему 'Modular som for dynamic object identification'

Modular som for dynamic object identification Текст научной статьи по специальности «Компьютерные и информационные науки»

CC BY
150
66
i Надоели баннеры? Вы всегда можете отключить рекламу.
Ключевые слова
NEURAL NETWORKS / MODULAR NEURAL NETWORKS / FORECASTING / TIME SERIES PREDICTION / DYNAMIC OBJECT IDENTIFICATION / VQTAM / RSOM

Аннотация научной статьи по компьютерным и информационным наукам, автор научной работы — Averkin A. N., Povidalo I. S.

Идентификация динамического объекта затруднена, если в исходных данных присутствуют шумы, некоторые параметры динамического объекта изменяются по неизвестным законам или точное число параметров объекта не известно. В таких случаях для решения задачи идентификации могут быть успешно применены нейронные сети. Существует множество различных типов нейронных сетей, применимых для идентификации динамического объекта. Среди всевозможных нейросетевых структур, применимых для решения задачи идентификации, выделяется класс нейронных сетей, основанных на самоорганизующихся картах (SOM). В данной статье описан ряд нейронных сетей, основанных на самоорганизующихся картах и успешно применимых для идентификации динамических объектов. Представлены уникальные модульные нейронные сети на основе SOM, полученные в результате исследований коры головного мозга млекопитающих, где в качестве модулей применены векторная квантовая память и рекуррентные самоорганизующиеся карты. Подробно описаны структура и алгоритмы обучения и работы таких сетей, также приведены некоторые результаты экспериментов и сравнения с другими нейросетевыми алгоритмами идентификации динамических объектов.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Object identification is complicated if noises are present in the source data, some of the object parameters change according to unknown laws or the exact number of the object parameters is unknown. In such cases neural network can be applied for dynamic object identification. There are a lot of different types of neural networks that can be used for dy-namic object identification. Among different neural network architectures applicable for dynamic object identification a class of neural networks based on self-organizing maps (SOM) can be noted. In this article, a number of neural networks based on self-organizing maps that can be successfully used for dynamic object identification are described. Unique SOM-based mod-ular neural networks, inspired by mammal''s brain cortex studies, with vector quantized associative memory and recurrent self-organizing maps as modules are presented. The structure and algorithms of learning and operation of such SOM-based neural networks are described in details. Some experimental results and comparison with some other neural networks are giv-en.

Текст научной работы на тему «Modular som for dynamic object identification»

MODULAR SOM FOR DYNAMIC OBJECT IDENTIFICATION

A.N. Auerkin, Ph.D. (Physics and Mathematics), Associate Professor (Dorodnicyn Computing Centre, RAS, Vavilova St. 40, Moscow, 119333, Russian Federation,

averkin2003@nbox. ru); I. S. Povidalo, Postgraduate Student (Dubna Internacional University for Nature, Socitty and Man, Universitetskaya St. 19, Dubna, 141980, Russian Federation, ipovidalo@gmail.com)

Received 19.05.2014

Abstract. Object identification is complicated if noises are present in the source data, some of the object parameters change according to unknown laws or the exact number of the object parameters is unknown. In such cases neural network can be applied for dynamic object identification. There are a lot of different types of neural networks that can be used for dynamic object identification. Among different neural network architectures applicable for dynamic object identification a class of neural networks based on self-organizing maps (SOM) can be noted. In this article, a number of neural networks based on self-organizing maps that can be successfully used for dynamic object identification are described. Unique SOM-based modular neural networks, inspired by mammal's brain cortex studies, with vector quantized associative memory and recurrent self-organizing maps as modules are presented. The structure and algorithms of learning and operation of such SOM-based neural networks are described in details. Some experimental results and comparison with some other neural networks are given.

Keywords: neural networks, modular neural networks, forecasting, time series prediction, dynamic object identification, VQTAM, RSOM.

УДК 004.9 Дата подачи статьи: 19.05.2014

ИДЕНТИФИКАЦИЯ ДИНАМИЧЕСКИХ ОБЪЕКТОВ С ПОМОЩЬЮ МОДУЛЬНЫХ САМООРГАНИЗУЮЩИХСЯ КАРТ

Аверкин АН., к.ф.-м.н., доцент (Вычислительный центр им. А.А. Дородницына РАН, ул. Вавилова, 40, г. Москва, 119333, Россия, averkin2003@inbox.ru);

Повидало И С., аспирант (Международный университет природы, общества и человека «Дубна», ул. Университетская, 19, г. Дубна, 141980, Россия, ipovidalo@gmail.com)

Аннотация. Идентификация динамического объекта затруднена, если в исходных данных присутствуют шумы, некоторые параметры динамического объекта изменяются по неизвестным законам или точное число параметров объекта не известно. В таких случаях для решения задачи идентификации могут быть успешно применены нейронные сети. Существует множество различных типов нейронных сетей, применимых для идентификации динамического объекта. Среди всевозможных нейросетевых структур, применимых для решения задачи идентификации, выделяется класс нейронных сетей, основанных на самоорганизующихся картах (SOM). В данной статье описан ряд нейронных сетей, основанных на самоорганизующихся картах и успешно применимых для идентификации динамических объектов. Представлены уникальные модульные нейронные сети на основе SOM, полученные в результате исследований коры головного мозга млекопитающих, где в качестве модулей применены векторная квантовая память и рекуррентные самоорганизующиеся карты. Подробно описаны структура и алгоритмы обучения и работы таких сетей, также приведены некоторые результаты экспериментов и сравнения с другими нейросетевыми алгоритмами идентификации динамических объектов.

Ключевые слова: нейронные сети, модульные нейронные сети, прогнозирование, предсказание временных рядов, идентификация динамических объектов, VQTAM, RSOM.

Identification theory solves problems of constructing mathematical models of dynamic systems according to the observations of their behaviour. The object identification step is one of the most important steps while constructing mathematical models of objects or processes. The quality of the model relies on this step. Therefore, the quality of control that is based on this model or results of a research with this model also rely on this step.

Dynamic object identification is one of the basic problems which can be solved using many different methods. For example statistic analysis or neural networks [1] can be used. Object identification is complicated if noises are present in the source data, some of the object parameters change according to unknown laws or the exact number of the object parameters is unknown. In such cases neural network can be applied for dynamic object identification. There are a lot of different types of neural networks that can be used for dynamic object identification.

Among different neural network architectures applicable for dynamic object identification a class of neural networks based on self-organizing maps (SOM) can be noted. Neural networks of such type will be given special attention in this article due to their wider spread and successful application in solving different kinds of problems of recognition [2], identification [3], etc. A number of biomorphic neural networks, architecture of which is the result of studying the structure of the cerebral cortex of mammals, will also be considered.

SOM-based neural networks can be used for dynamic object identification

Problem definition. Identification of a dynamic object that receives a vector of input parameters u(t) and has a corresponding output vector y(t) can be described as finding the type of a model of this object, that has an output y(t) and finding parameters of this

model such that minimize error e = || y(t) - y(t)||2 of this model (see fig. 1).

u (t) Dynamic у (t)

object i 1

i r

1 \

* ^ Model у(t) T

e-> 0

-►

Fig. 1. Dynamic object identification scheme

Suppose that there is a vector of input parameters u(t), t = 0,T -1 and a sequence of output vectors y(t),

t = 0, T -1, where T is a number of input-output pairs. We will consider the solution of the object identification problem as a definition of the function type f that will define the model of identified object: y(t) = f [y(t -1),..., y (t - n), u(t), u (t -1),..., u (t - n)], where y(t) is a vector of output parameters of the model. At a single moment of time the input of the model takes current known measured values of the input parameters along with n„«T previous values and ny<<T previous output parameters of the identified object.

Also we will consider solution of this problem in the following form:

y (t) = f [y(t -1),..., y (t - ny), u(t), u(t -1),..., u(t - n„)], where y(t) = y(0), y(t) is a vector of output parameters of the model. This identification scheme has recurrent connections and at a single moment of time the input of the model takes current known measured values of the input parameters of the object along with nu<<T previous values of these parameters and ny<<T, previous output parameters of the model.

Vector quantized temporal associative memory (VQTAM). VQTAM is a modification of self-organizing maps which can be used for identification of dynamic objects [4, 5]. The structure of this network is similar to regular SOM structure. The key difference is in the way that weight vectors are organized. The input vector x(t) of this network is split into two parts: x'"(t), xoui(t). The first part of the input vector x'"(t) contains information about the inputs of the dynamic object and its outputs at previous time steps. The second part of the input vector xoui(t) contains information about the expected output of the dynamic object corresponding to the input x'"(t). The weights vector is also split into two parts in a similar

way [5]. Thus x(t) =

n (t ) " (t)

and w, (t) =

w

(t) (t)

where x,(t) is weights vector of the i-th neuron,

w" (t) is the part of the weights vector that contains information on the inputs corresponding to x1 n(t), and w°"' (t) is the part of the weights vector that contains information on the outputs corresponding to xoui(t). The first part of the input vector contains information on the process inputs and its previous outputs: xin (t) = (y (t-1),..., y (t - n), u (t), u(t-1),..., u(t - n)), where ny<<T, n„«T. The second part of the input vector xoui(t)= y(t) contains information on the expected output of the process corresponding to the inputs xin(t).

Each vector in the learning sample consists of a pair of vectors (y(t), u (t)) and the sample should contain not less than max(nu, ny) vectors. Vectors y(t) are the measured output parameters of the process at time step t and u(t) are the input parameters of the process at the same time step.

After presenting a subsequent input vector x(t), combined of several vectors from the learning sample, to the network the winner neuron is determined only by the x'"(t) part of the vector:

i* (t) = arg min { x" (t) - wf (t )|| J,

where i*(t) is a number of the winner neuron at time step t.

For weight modification a modified SOM weight modification rule is used:

Awf (t) = a(t)h(i*, i, t)[xin (t) - wf (t)],

Awf(t) = a(t)h(i*,i,t)[xou'(t) - w°u'(t)],

where 0<a(t)<1 is a learning rate of the network, h is a neighbourhood function of the i-th and i*-th neurons. For example, a Gaussian function can be chosen as a neighbourhood function h(i*, i, t):

h(i*, i, t ) = exp

||r (t) - r* (t )|| 2a2 (t)

||2 Л

v y

where r,(t) and r,»(t) are positions on the map of the ith and i*-th neurons, u(t)>0 determines the radius of the neighbourhood function at time step t. When the winner neuron i* is defined, the output of the network is set to (t) (learning scheme of the network is

presented on fig. 2).

On the test sample VQTAM's input takes only x1 n(t) part of the input, which is used to define the

x

x

v Wi

u(t)

У (t -1)

b-l Q

н

x"" (t )

Serial mode

о >

wo1" (t ) = y(t )

U (t)

h-l Q

H

x"" (t )

О >

Parallel mode

wou' (t ) = y(t )

Fig. 3. VQTAM's working scheme

winner neuron, and the output of the network is set to wO" (t ) (the working scheme is presented on fig. 3).

Vector wO" (t) can be interpreted as the predicted

output y(t) of the dynamic object at time step t. It is worth mentioning that this network can be successfully used to identify objects with both continuous and non-continuous outputs.

Recurrent self-organizing map (RSOM). In RSOM unlike conventional SOMs with recurrent connections, a decaying in time vector of outputs is introduced for each neuron. This vector is used to determine the winner neuron and in maps weights modification [6, 7].

The network inputs vector x(t) is represented as follows:

x(t) = (y (t-1),..., y (t - ny ), u (t ), u (t -1),..., u (t - n )), where ny<<T, nu<<T.

Output of each neuron is determined according to the following equation: V (t) = || v, (t)||, where

v,(t)=(1-a)v,(t-1)+a(x(t)-w,(t)), a is the output decay factor (0<a<1), V,(t) is the output of the i-th neuron at time step t, w,(t) is weights vector of the i-th neuron, i = 1, k, k is the neuron count.

After presenting a subsequent input vector to the network a winner neuron is determined as the neuron with minimal output [7]: i* (t) = argmin {Vi (t )}.

To modify the weights a modified conventional SOM rule is used: Aw, (t) = a (t)h(i*, i, t)v i (t ), where

0<a(t)<1 is learning rate, h is a neighbourhood function of i-th and i -th neurons. The neuron schematic of this network is shown in figure 4.

When the learning process is complete, the network is presented again with the learning sample and clusterizes it. Each cluster can be approximated with an individual model, for example a linear function fi(t) for the i-th cluster. Thus, after presenting the learning sample a linear function is defined for each vector of this sample. These functions can be used to determine the output value at the next time step.

This process can be speeded up using algorithms for constructing local linear models while training the neural network. Such method is applicable in case if the size of the output is scalar, thus >>(t) and y(t) are scalar. This method is applicable only to the objects with outputs that can be described with continuous functions. Each neuron of the RSOM network is associated with a matrix 4(0 that contains coefficients of the corresponding linear model:

4(t) = [by (t),...,bI>; (t),a,, (t),...,a,^ (t)J .

The output value of the network is defined according to the following equation:

n %

Ki) = S b, k (t)u(t - k) + £ a

k=1 ' 1=1

,(t)y(t -1) = 4 (t)x(t),

where 4,.(t) is a matrix of coefficients associated with the winner neuron i*(t). Matrix 4,.(t) is used for linear approximation of model output.

While constructing the local linear models simultaneously with the training of the neural network an additional rule to modify the coefficients of the linear model is needed: 4 (t +1) = 4 (t) + a'h(i*, i,t)A4 (t), where 0<a'<1 is a learning rate of the model, A4,(t) is Widrow-Hoffs rule for error correction:

M (t) = [y(t) - A (t)x(t)]■

x(t)

I ll2

x(t )

where >>(t) is desired output of the model for the x(t) input.

Thus, at each step of the network training a modification of the model coefficients is performed along with modification of the weights of neurons. On the test sample after an input vector is presented to the network a winner neuron i*(t) is chosen. Than a corresponding coefficients matrix 4,»(t) of the linear model is calculated. Using the determined matrix the output

z

of the model is defined by the equation: y(t) = A* (t)x(t). The working scheme is shown on figure 5.

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.

As soon as the winner module is defined, the adaptation process takes place and the weights of the module are being modified according to one of learn-

u (t)

У (t -1)

b-l Q

H

x(t )

О от Pi

u(t)

A* (f ) ^s

У (t )

b-l Q

H

x(t )

О от Pi

4(t y(t )

Serial mode

Parallel mode

Fig. 5. RSOM's working scheme

Modular self-organizing maps. Modular self-organizing maps are presented in Tetsuo Furukawa's works [8, 9]. Modular SOM has a structure of an array which consists of functional modules that are actually trainable neural networks, such as multilayer perceptrons (MLP), but not a vector, as in conventional self-organizing maps. In case of MLP-modules modular self-organizing map finds features or correlations in input and output values simultaneously building a map of their similarity. Thus, a modular self-organizing map with MLP modules is a self-organizing map in a function space but not in a vector space [9].

These neural network structures can be considered biomorphic, as their emergence is due to research of the brain structure of mammals [10], and confirmed by a number of further studies [11]. The basis of the idea of the cerebral cortex structure is a model of cellular structure, where each cell is a collection of neurons, a neural column. Columns of neurons are combined in more complex structures. In this regard it was suggested to model the individual neural columns with neural networks [11]. This idea has formed the basis of the modular neural networks.

In fact, the modular self-organizing map is a common SOM, where neurons are replaced by more complex and autonomous entities such as other neural networks. Such replacement requires a slight modification of the learning algorithm.

In the algorithm proposed by Furukawa [9] at the initial stage the network receives the i-th sample of the input data corresponding to I functions which will be mapped by the network, and the error is calculated

for each network module:

1 J

E = - Ell

||y i, j y i, j ,

where k is a module number for which the error is calculated, J is number of vectors in the sample, yk j

is an output of the k-th module, y,,,- is expected output of the network on suggested set of input data. Winner module is calculated as the module that minimizes the error E\: k* = argmin Ef.

ing algorithms suitable for the networks of this type. After that the weights of the main SOM are being modified. In this process parameters (weights) of each module are considered as the weights of the SOM and are modified according to standard learning algorithms suitable for conventional SOMs.

In this study SOMxVQTAM and SOMxRSOM networks were developed. They are SOMs with modules of VQTAM type and SOMs with modules of RSOM type respectively. Further some application results of such networks will be reviewed.

Modular network where VQTAM networks are used as modules (SOMxVQTAM) that was developed during this work is trained by combination of the learning algorithm of modular SOMs and VQTAM learning algorithm. When a subsequent learning sample vector is presented to the network, outputs of all VQTAM networks (that are used as modules) are calculated. Then a winner-module is selected as the module with the output with less deviation from the expected output for the presented sample vector. Weights of the winner-module network are modified according to the VQTAM learning algorithm. The set of weight vectors of the whole module-network is treated as one of the weight vectors of the whole modular network, and the weight modification is done according to the standard f SOM learning algorithms.

In case of a modular network, where RSOM networks are used as modules (SOMxRSOM) also developed during this work, learning is held according to similar algorithm.

Using SOM-based neural networks for dynamic object identification

For some experiments and comparisons of the algorithms the neural networks of types VQTAM, RSOM, SOMxVQTAM and SOMxRSOM were tested on samples that were used in 2008 to identify the winners at neural networks forecasting competition [12]. Results of this competition were used for this

z

2

study as there is a detailed description of the place definition method used for all competitors. Also a fair amount of different algorithms took part in this competition and there was a description for the most of those algorithms as well as the learning and testing samples which allowed a comparative analysis of the neural networks described in this paper with other advanced algorithms. To determine the place in the table of the competition organizers [12] it is suggested to forecast 56 steps for each of the 111 samples. For each of the resulting predictions a symmetric mean absolute percentage error SMAPE was calculated:

SMAPE =1 £

У (t ) - У (t )

-*100.

n '=1 (y(t) + y(t))/2

where >>(t) is a real state of the object at time step t, y(t) is an output of the model at time step t, n is a number of vectors in the testing sample (56 for this competition). Then the place in the table was determined from the average error for all of 111 samples.

Each of the 111 samples had different lengths; those samples represented a monthly measure of several macroeconomic indicators.

Results table for different forecasting algorithms

№ Algorithm name SMAPE, %

1 Wildi 19,9

2 Andrawis 20,4

3 Vogel 20,5

4 D'yakonov 20,6

5 Noncheva 21,1

6 Rauch 21,7

7 Luna 21,8

8 Lagoo 21,9

9 Wichard 22,1

10 Gao 22,3

— SOMxVQTAM 23,4

11 Puma-Villanueva 23,7

— VQTAM 23,9

12 Autobox (Reilly) 24,1

13 Lewicke 24,5

14 Brentnall 24,8

15 Dang 25,3

16 Pasero 25,3

17 Adeodato 25,3

— RSOM 25,7

18 undisclosed 26,8

19 undisclosed 27,3

- - SOMxRSOM 27,9

This and other tests that were concluded lead to a suggestion that modular modification gives a significant increase if accuracy in case of VQTAM modules, but modular networks are more sensitive to changes of learning parameters. The higher SMAPE in case of the SOMxRSOM modular network compared to RSOM network can be explained by the fact that RSOM itself contains local models and during the learning process of the SOMxRSOM network those

local models are treated as weight vectors for the whole SOMxRSOM network. But the local models are constructed for different parts of the input data and the local model corresponding to one of the neurons of one of the RSOMs are most likely to be constructed for the different part of the input data compared to the local model contained in another RSOM for the neuron on the same position.

Conclusion

In this article several SOM-based neural networks that can be successfully used for dynamic object identification have been described. New modular networks with VQTAM and RSOM networks as modules where developed. In case of the VQTAM modules a significant increase in quality of identification is achieved in comparison to regular VQTAM networks. In case of neural networks where RSOMs are used as modules quality of the identification is decreased that can be explained by the features of algorithm for constructing linear models. This results show that it is possible to increase the quality of the identification by simply joining neural networks in to modular structures.

References

1. Haykin S. Neural Networks — a comprehensive foundation. 2nd ed. Prentice Hall Int., Inc., 1998, 842 p.

2. Efremova N., Asakura N., Inui T. Natural object recognition with the view-invariant neural network. Proc. of 5th int. conf. of Cognitive Science (CoSci). 2012, June 18-24, Kaliningrad, Russia, pp. 802-804.

3. Trofimov A., Povidalo I., Chernetsov S. Usage the self-learning neural networks for the blood glucose level of patients with diabetes mellitus type 1 identification. Science and education. 2010, no. 5, available at: http://technomag.edu.ru/doc/142908.html (accessed July 10, 2014) (in Russ.).

4. Souza L.G.M., Barreto G.A. Multiple local ARX modeling for system identification using the self-organizing map. Proc. of European Symp. on Artificial Neural Networks — Computational Intelligence and Machine Learning. Bruges, 2010.

5. Koskela T. Neural network methods in analyzing and modelling time varying processes. Espoo, 2003, pp. 1-72.

6. Varsta M., Heikkonen J. A recurrent Self-Organizing Map for temporal sequence processing. Springer Publ., 1997, pp. 421-426.

7. Lotfi A., Garibaldi J. Applications and Science in Soft Computing. Advances in Soft Computing Series, vol. 24, Springer Publ., 2003, pp. 3-8.

8. Tokunaga K., Furukawa T. SOM of SOMs. Neural Networks. 2009, vol. 22, pp. 463-478.

9. Tokunaga K., Furukawa T. Modular network SOM. Neural Networks. 2009, vol. 22, pp. 82-90.

10. Logothetis N.K., Pauls J., Poggiot T. Shape representation in the inferior temporal cortex of monkeys. Current Biology. 1995, vol. 5, pp. 552-563.

11. Vetter T., Hurlbert A., Poggio T. View-based Models of 3D Object Recognition: Invariance to imaging transformations. Cerebral Cortex. 1995, vol. 3, pp. 261-269.

12. Artificial Neural Network & Computational Intelligence Forecasting Competition. 2008. Available at: http://www.neural-forecasting-competition.com/NN5/results.htm (accessed July 10, 2014).

Литература

1. Haykin S. Neural Networks - a comprehensive foundation. 2nd ed. Prentice Hall Int., Inc., 1998, 842 p.

2. Efremova N., Asakura N., Inui T. Natural object recog-

nition with the view-invariant neural network. V Междунар. конф. по когнитивной науке: тез. докл. (18-24 июня 2012 г.), Калининград). 2012. С. 802-804.

3. Трофимов А.Г., Повидало И.С., Чернецов С.А. Использование самообучающихся нейронных сетей для идентификации уровня глюкозы в крови больных сахарным диабетом 1-го типа // Наука и образование. 2010. №. 5; URL: http://tech-nomag.edu.ru/doc/142908.html (дата обращения: 10.07.2014).

4. Souza L.G.M., Barreto G.A. Multiple local ARX modeling for system identification using the self-organizing map. Proc. of European Symp. on Artificial Neural Networks - Computational Intelligence and Machine Learning. Bruges, 2010.

5. Koskela T. Neural network methods in analyzing and modelling time varying processes. Espoo, 2003, pp. 1-72.

6. Varsta M., Heikkonen J. A recurrent Self-Organizing Map for temporal sequence processing. Springer Publ., 1997, pp. 421-426.

7. Lotfi A., Garibaldi J. Applications and Science in Soft Computing. Advances in Soft Computing Series, vol. 24, Springer Publ., 2003, pp. 3-8.

8. Tokunaga K., Furukawa T. SOM of SOMs. Neural Networks. 2009, vol. 22, pp. 463-478.

9. Tokunaga K., Furukawa T. Modular network SOM. Neural Networks. 2009, vol. 22, pp. 82-90.

10. Logothetis N.K., Pauls J., Poggiot T. Shape representation in the inferior temporal cortex of monkeys. Current Biology. 1995, vol. 5, pp. 552-563.

11. Vetter T., Hurlbert A., Poggio T. View-based Models of 3D Object Recognition: Invariance to imaging transformations. Cerebral Cortex. 1995, vol. 3, pp. 261-269.

12. Artificial Neural Network & Computational Intelligence Forecasting Competition. 2008. Available at: http://www.neural-fo-recasting-competition.com/NN5/results.htm (accessed July 10, 2014).

i Надоели баннеры? Вы всегда можете отключить рекламу.