Научная статья на тему 'An approach to reconstruct target function of the optimization problem with precedent initial information'

An approach to reconstruct target function of the optimization problem with precedent initial information Текст научной статьи по специальности «Компьютерные и информационные науки»

CC BY
23
11
i Надоели баннеры? Вы всегда можете отключить рекламу.

Аннотация научной статьи по компьютерным и информационным наукам, автор научной работы — Anafiyev Ayder, Abdulkhairov Alim

The optimization problem with precedent (training sample) initial information is considered. Some approaches for reconstruction of the target function of such optimization problem are proposed. The open problems that must be solved to obtain better quality solutions of this problem are highlighted.

i Надоели баннеры? Вы всегда можете отключить рекламу.
iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.

Текст научной работы на тему «An approach to reconstruct target function of the optimization problem with precedent initial information»

AN APPROACH TO RECONSTRUCT TARGET FUNCTION OF THE OPTIMIZATION PROBLEM WITH PRECEDENT INITIAL

INFORMATION © Ayder Anafiyev, Alim Abdulkhairov

Taurida National V. I. Vernadsky University Mathematics and Computer Science Department e-mail: anafiyev@gmail.com, alim.abdulkhairov@gmail.com

Abstract. The optimization problem with precedent (training sample) initial information is considered. Some approaches for reconstruction of the target function of such optimization problem are proposed. The open problems that must be solved to obtain better quality solutions of this problem are highlighted.

1. Formulation of the problem

Let X, Y and W are the spaces of object (feature), target function value and admissible function value respectively, f : X ^ Y is a target function and Q is an admissible set of objects. Consider the optimization problem

f (x) —max (1)

J y ' X e n c x w

with initial information is represented by the set of triple X^ = (xi,yi,wi)ii=l, where xi E X, yi E Y and wi E W. The triple (xi,yi,wi) will be called precedent (training sample). If W = {0,1} then wi = 0 means that the object xi E Q, otherwise (wi = 1) means that the object xi E Q. If W = [0,1] then wi E W could be interpreted as the probability that the object xi belongs to the set Q.

The problem (1) will be called the optimization problem with precedent initial information [1, 2, 3]. This is a problem with incomplete information. For solving this problem it's necessary to construct an algorithm which finds in Q the optimal object(s) of the target function f or reduces the problem to a certain optimization problem with a fully defined data and which allows to find an effective decision.

The optimization problem could be divided into two problems: the problem of reconstructing of the target function f (regression problem) and the problem of reconstructing the admissible object set Q (classification problem). There are many approaches for solving the regression and classification problems. However it's still open the problem of synthesis of these two problems to get the better quality solution of the given optimization problem.

Example 1. Lets consider the maximization problem with precedent initial information: X = r2, Y = r, W = {0,1}. The training sample Xe is set as a training table:

x1 x2 y w

2 2 2 1

5 1 3 1

4 4 4 1

6 5 - 0

8 4 - 0

10 5 - 0

Admissible objects are marked by circles, inadmissible — by crosses. In additional the objects with known values are labeled by the target function values.

As you can see from the figure when we come near to the imaginary border of the space Q the value of the target function grows. Obviously this information would be useful for the decision making. The location of objects of different classes ("admissible" and "inadmissible" objects) may be very important during the target function reconstruction.

2. Reconstruction of the target function

2.1. Linear regression. Lets consider an input space X = rn and output space Y = r. The linear regression model $(x,a) is represented by

0{x,a) =

aj g r, j = i, n.

i= 1

The optimal value of parameter a is selected from solution of the optimization problem

*

a =

arg min y L(a,xi) = arg min Q(a, Xe)

i=1

where L(a,x) is a loss function which is used to determine loss on the object x and

e

Q(a, Xe) = Y, L(a , xi) is an empirical risk.

i= 1

When defining the loss function it must be considered the fact that we reconstruct the target function of the maximization problem (1). In this case we need more "detailed" study the objects on which the target function takes large values. Therefore the object importance depends on the value of the target function for this object. Thus the loss function will be considered as

Ly (a,x) = y (x)L(a,x)

where y(x) is the weight function which defines an importance of the object x for the optimization problem.

j

Let us consider L(a,x) = (0(a,x) — f (x))2 and y(x) = f (x).

Introduce the matrix notations F = (xj y = (yi)exi, a = (a )nx1

and 7 = diag(Y(xi),... ,7(x^)) = diag(yi,... ,ye).

Let us write the optimization problem (2) in matrix form

Q(a,X*) = y||Fa — y\\2 ^ min .

a

The standard way to solve this optimization problem is to use a necessary condition of minimum

dQ = 2FT y (Fa — y) = 0.

Therefore

FT YFa = FT Yy.

If FtyF is a nonsingular matrix then the solution of the system will

a* = (FTYF)-1FTYy.

The result of the using the linear regression method for reconstruction of the target function f of the optimization problem (1) is illustrated on the figure 1. As you can see a more detailed learning of the optimal objects could improve the solution1.

Fig. 1. Application of linear regression method for reconstruction of a target function of an optimization problem (bold line — with proposed loss function)

1Of course this approach requires more detailed research.

2.2. Support Vector Regression. The SVR [4] (Support Vector Regression) could be used to solve the problem of reconstructing the target function.The SVM (Support Vector Machines) method is used for the reconstruction of the admissible object set Q. The support vector machines is one of the best classification algorithm nowadays. Learning of SVM leads to solving a quadratic (linear) programming problem. The position of the discriminant hyperplane depends only from a few support objects. In addition the use of kernel functions allows the efficient using this method for both: linearly separable and inseparable samples.

The figure 2 shows an example of the optimization problem with the precedent initial information with six objects: two are inadmissible and marked with a cross symbol and the other four belong to the space Q and marked as a circle (the big radius circles have the larger target function value).

Fig. 2. SVR and SVM for solving the optimization problem with precedent initia,! information.

Line L1 is the result of the solving regression problem and corresponds to the target function. Line L2 is the result of solving classification problem on two classes: "admissible" and "inadmissible" objects. The set D is very interesting from scientific point and needs to be researched.

Consider the applying of SVR method to the reconstruction of the target function of the optimization problem 1. The target function f (x) is represented as

f (x) = (a,x) + ao,

where (•, •) is the scalar product.

The optimal set of parameters (a^,..., an, a0) is the solution of the optimization problem

0

1,,

-a

2 il I

+ C ^^ LY(a, Xj)

i=1

a, xj ^ mm,

L7(a,Xj) ^ e + ^ 0, i = 1,1. where C is an error on the object x,.

It's proposed to use LY(a,x) as a loss function

L£)7 (a, x)

0, y(x)L(a, x) ^ e

Y(x)L(a,x) — e, y(x)L(a,x) > e,

where y(x) is the weight function which defines the importance of the object x for optimization problem (1).

Let us &x = Y(x)L(a,x) — e. Then

L£;7 (a, x)

0, £* ^ 0, Cx, Cx > 0 ,

and

Leri (a,x) = i( |&x| + Cx) •

Introduce additional variables C+ and £,

= Mxl+ix, c = ^, s 0, Cx- > 0.

Note that

2

Cx Cx Cx 1 Cx1 Cx + Cx

As a result the optimization problem is got in the form below

1.. . 2 ||a|

+ C£C;

i=1

+ ^ min, 1 a,ao

Y(Xi)L(a,Xi) ^ £ + £+ - Cxi,

C+ ^ 0, Cx- ^ 0, i = M,

which reduces to the problem of quadratic (linear) programming.

2

Conclusion

The optimization problem with precedent (training sample) initial information is considered. The open problems that must be solved to obtain better quality solutions of this problem are highlighted. Proposed the weighted loss function which uses the importance of the object for the optimization problem. It's shown how to use such function for reconstruction of the target function using linear regression and SVR methods. It should be noted that the using of different weight loss functions (with weights depending on an optimization problem) provide more accuracy formalize the optimization problem and obtain better solutions.

References

1. Anafiyev, A. and Blyschik, V. 2011. Optimization problem with precedent initial information. Taurida Journal of Computer Science Theory and Mathematics, 2, pp. 51-57.

2. Anafiyev, A. 2012. The approach to solving optimization problems with precedent initial information. Taurida Journal of Computer Science Theory and Mathematics, 2, pp. 7-12.

3. Taratynova, N. 2006. Building an optimization model for precedent initial information as a problem of nonlinear regression. Iskusstvennyi intellekt, 2, pp. 238-241.

4. Smola, A. 2004. Tutorial on Support Vector Regression. Statistics and Computing, 14, pp. 199-222.

iНе можете найти то, что вам нужно? Попробуйте сервис подбора литературы.
i Надоели баннеры? Вы всегда можете отключить рекламу.