METHOD OF ACCOMPANYING SEPARATION (PORA) [1,2].

     If we will construct the optimal separation border between patterns i and j then it will be possible to find, that this border at the same time separates another couples of patterns with an error R, not exceeding the permitted value. For example, the solving border (see Fig.1) created for separation of patterns 2 and 5 may as well be used for separation of patterns 1 and 4 from patterns 3 and 6. 
 

 
Fig.1 
     The process must be started from separation of the two most hardly separated patterns. All couples of patterns, successfully separated by the created border, are marked. This couples are excluded from the couples list. The next most hardly separated patterns are chosen from unseparated couples and the next separating border is created for this couple. Such steps are repeated until the complete exhausting of unseparated couples list. 
     For successful position of patterns the number of separating borders h may be much lower than the number of patterns to be recognised. The list of k1 patterns, lying to one side from the border, and k2 patterns, lying to another side, is fixed for every border. Let us characterise the «informativeness» of the border by the value I=1-|k1-k2|/k ,which is in the range from 0 to 1 and is equal to 1, if the border separates the pattern multitude into two equal parts. 
     The recognition process, applying the created borders, is the following. Starting from the most informative border, it is discovered – to which side of the border the control point q is situated and all patterns situated to the other side from the border are excluded from the pretenders list. On the base of the following informative border the list of pretenders is shortened again. Such procedure is repeated until the list will contain the only one pattern corresponding to recognition point q. 
 
METHOD OF BY-COORDINATE EXCLUSION [2,3].

     Let us imagine that we consider the losses of «aim missing» type equal to R0 as admissible. 
    Such losses may occur, when the separation border lies at the distance d between the control point q and mathematical expectation of i-th pattern, equal to *i. If the control point q is remotes from math expectation of i-th pattern by the distance higher than d, then we can think that it does not belong to pattern i. 
     Every coordinate of properties space makes its own contribution and, if at least one l-th coordinate gives the distance r(ql,*il) for pattern i higher or equal to d, then i-th pattern may be excluded the list of competitors. 
     This consideration creates the basis of by-coordinate exclusion method, which consists in the following. The projections of point q and distribution of all patterns at every coordinate separately are examined. The distances r between the point ql and mathematical expectations *il for all k patterns are defined for each l-th coordinate. The patterns, satisfying the condition r(ql,*il)*d are excluded from the list of patterns, pretending to include the point q. The same procedure is repeated for remaining patterns with application of projection to the second coordinate and so on until the moment, when list of pretenders will contain only the fixed number k1 of patterns. For these most strong pretenders the calculation of distances and estimation of expected losses is performed in n-dimension space with using of optimal decision rules. 
 

Method " Splitting of the standads" (algorithm DRET).

    This method reflects aspiration to correct recognition of training sample, but with use of coverings of training sat of each pattern by simple figures, which are becoming complicated as required [3,4].  
    One of variants of this method provides use as covering figures of a set hypersphere. For each of k of patterns the sphere of minimal radius covering all its training realizations is under construction. Value of radiuses of these spheres and distances between their centers allows defining patterns, which spheres are not crossed with spheres of other patterns. It is considered such spheres reference (see pattern 1 in a fig. 2), and their centers and radiuses are remembered in quality   "of the standards of the first generation ". 

 
 
Fig. 2.

    If two patterns are crossed, but in the field of crossing there was no realization of training sat, such crossing is considered fictitious, the centers and radiuses of these spheres as are brought in to the list of the standards of the first generation. Thus the area of crossing is considered to belonging sphere with smaller radius (sphere 2 in a fig. 2).  
    If in a zone of crossing there were points only of one pattern, this zone is considered belonging to this pattern (sphere 3 in a fig. 2). The point will be considered concerning to an pattern 4, if it gets in sphere 4 and does not get in sphere 3.  
    If the area of crossing contains points of different patterns, for these points the standards of the "second generation " (sphere 4,5 and 6) are under construction. If also they are crossed, for points from a zone of crossing the standards of the "third generation" are under construction. The procedure of splitting of the standards proceeds before reception of given reliability of recognition of a training sequence.  
    The experience has shown, what even in very complex cases for good recognition of training sat it is enough on the average no more than three generations of the standards.  
    At recognition of new realizations the process of check begins with spheres of the minimal diameter.  
 

Taxonomic decision functions (algorithm ÒDF) [3,4].

    Usually on a grade level the decisive rule is under construction, which parameters are defined and are rigidly fixed under the information contained in training sat. What control realizations would not be showed then for recognition a decisive rule does not vary. But other approach is possible also: a decisive rule to build directly during recognition, leaning on the information contained as in the trainee, and in control sats.  
    In a fig. 3. objects of training sat of two patterns are designated by points and by crosses, and the realizations of control sat are designated by small points. If to approximate clots of points of training sample by unimodal distributions, the decisive border will represent a line a. Such decisive rule constructed on not representative sample, will give many mistakes at recognition of control sample.  
 

 
 
Fig. 3.

    If training and control sat to consider in common, it is possible to construct decisive border as a line b, which distinguishes training sat also well, as well as the line a, but reduces quantity of mistakes on control sat. Such decision turns out if to apply criteria taxonomy: the large distances between taxons and small distances inside taxons. It also is necessary in a basis of acceptance of the decisions by a method of the taxonomic decision functions.  
    Mix of training and control objects are exposed taxonomy on k taxons with the help of algorithms such as FOREL or KRAB. If in some taxon there are points of training sample only one i-th of an pattern, all control points which have got in this taxon, concern also to an pattern i. If in taxon there are points from k ' of different patterns, such taxon is broken on k ' finer taxons. This procedure proceeds so long as in everyone taxon there will be no training points only of one pattern. The control points, which have got in taxon, not containing of training points, are considered belonging to new, (k+1)-th pattern. At desire they can be attributed to one of k of patterns, the connection to which will give the greatest value of criterion of taxonomy quality.  
    The described algorithm allows to use not only information from training sat, but also from control. It provides higher stability of algorithm ÒRF to such frequently to meeting unpleasant phenomenon, as small of training sample.  
 

Algorithm DW [5].

    This algorithm concerns to a class " of Logic Decisive Rules " [6]. In its basis the hypothesis of "projective compactness " lays. Let's consider its work on an example of recognition of two patterns.  
    As the elementary logic statement J in this algorithm the expression of the following kind refers to as:  
  à) X (a) =x        for an attribute X, measured in a nominal scale.  
  á) X (a) < =x    for an attribute in any stronger scale.  
In these expressions x - fixed ("threshold") value. Let's name the statements X (a) = /x and X (a) > x additional to above-stated and to designate through J-. As conjunction of length l on the elementary statements we shall name expression of a kind  
S = J (1) & J (2) &... &J(l). 
    Set conjunctions we shall represent as a dichotomizing tree D, in which to each branch there corresponds one conjunction. The final tops of a tree contain names of objects of training sat, past here from initial top of a tree. If in some final top there are objects of different patterns, it is considered, that this top belongs to that pattern, whose objects in it are more than others.  
    The work on check of the validity or falsity of expressions S in relation to object q is organized with the help of the table T (i, j). To top of a tree the line of the table is compared i-th to number i (i.e. statement J), in which the characteristics j of this top are described:  
Ò (i, 1) - number of a lines occupied by the given top,  
Ò (i, 2) - number of a line, where it is necessary to pass, if J is true,  
Ò (i, 3) - number of a line, where it is necessary to pass, if J is false,  
Ò (i, 4) - number of an attribute, on which is constructed the statement J,  
Ò (i, 5) - threshold x, used at construction of the statement J,  
Ò (i, 6) - quantity of objects of the first pattern in i-that top,  
Ò (i, 7) - quantity of objects of the second pattern in i-that top.  
    The final tops of a tree will be marked with that at them T (i, 4) =0.  
    The movement on a tree with the help of the table T (i, j) is carried out as follows. Let we are in top i. If T (i, 4) = /0, whether that we check satisfies object q to conditions of the statement J, reflected in T (i, 4) and T (i, 5). If yes, we go in top T(i,2), if is not present - in top T(i, 3). If the next top is final, contents of its elements T (i, 6) and T (i, 7) are checked. If will appear, that T (i, 6) > T (i, 7), the object q will be recognized as realization of the first pattern and on the contrary. If T (i, 6) = T (i, 7), the decision is accepted for the benefit of this or that pattern with probability 0,5.  
    At a choice of variants of threshold value x and their combinations arises large testing, which reduction is based on a method of escalating " best to best ". On a first step best in sense of some criterion G the elementary statement J (11) and its addition J (12) = J (11-) is under construction. With the help of these statements the training sat is divided into two groups: group of objects U (11), satisfying J (11), and group U (12), satisfying J (12). If the group U (11) contains objects of different patterns, for it the statement J (21) and its addition J (22), best from the point of view of criterion G is searched. The same procedure for group U (12) gives the best statement J (23) and its addition J (24). In result the training sat is divided into four groups. The process of such division can be presented as a tree D, submitted on a fig. 4.  

                                                          / ----- J (21) - [70i, 3j]  
                                                        / 
                                            / ----- J (11) - [70i, 20j]  
                                          /             \ 
                                        /                 \ ----- J (22) - [0i, 17j] 
                                      / 
                        D - ---- J (1) - [100i, 100j]  
                                     \ 
                                       \                 / ----- J (23) - [20i, 75j] 
                                         \             /  
                                           \ ----- J (12) - [30i, 80j] 
                                                       \  
                                                         \ ----- J (24) - [10i, 5j] 

                                                           Fig. 4. 

     The question on, whether it is necessary further to continue process of escalating of branches, is decided by check of contents of each of the received groups. If in group U will find out h1 of objects of 1-st pattern and h2 of objects of the second pattern and min (h1, h2) > = f*m, the division of group U on subgroups should be continued. Here f - positive constant, smaller 1, and m - common number of objects of training sample. Otherwise group U forms final top and new object getting in it, concerns to a that pattern, whose training realizations it is more in this top than.  
     If to accept f=0,03, at m=200 it is required to make the further division only for top J (23).  
    The logic decisive rules have appeared a very effective means of the decision of tasks of recognition. They can work with different scale of attributes. It is not terrible, if any attribute at new object is not known: can appear, that to available attributes it satisfies to the certain laws and is well distinguished.  
    The process of construction LDF is long, but the process of acceptance of the decisions by the found rules is very simple and can be done even manually.  
    The large importance there is an evident and easily interpretive form of representation LDF as the list of conjunctions " If..., then... ". As has shown experience, some customers, having received  LDF, admit, that it is more interesting to them and it is useful, than an opportunity to use it for automatically recognition of new objects. To them become visible and the simple laws are clear which take place in area, investigated by them.  
    At last, the most important advantage LDF is, that the formulation of laws as conjunction corresponds to the form of representation of knowledge in intellectual systems. Hence, the methods of search LDF can be used for automatically extraction of knowledge from the data.  


REFERENCES:

1. Zagoruiko N.G. Linear decision functions close to optimum. In:  "Computing systems". -Vol. 19, Novosibirsk, 1965. 
2. Zagoruiko N.G. The combined methods of the decision making. In:  "Computing systems". - Vol. 24, Novosibirsk, 1966.  
3. Zagoruiko N.G. Methods of recognition and their applications. Published by "Soviet Radio". - Ì. - 1972. - 308 p. 
4. Zagoruiko N.G., Elkina V.N., Lbov G.S., Emelianov S.V. A package of the applied programs OTEKS. M.: Finance and Statistics, 1986. - 160 p. 
5. Manohin A.N. Methods of recognition of patterns based on logic decision functions. -In: An empirical prediction and recognition of patterns. (Computing systems, Vol. 67), Novosibirsk, 1976, p. 42-53. 
6. Lbov G.S. Methods of processing of different scale experimental data. Published by "Science" the Siberian branch, Novosibirsk, 1981. 160 p.