Академический Документы
Профессиональный Документы
Культура Документы
ISSN No:-2456-2165
Abstract:- In this note, we propose a new approach for determined by their evaluation tables. More specifically, we
benchmarking computational problems and their present data for benchmarking in the form of the
solvers. The proposed methods are based on special benchmarking context, i.e., as a triple S , P, J , where
construction of the paired comparison matrices of the
problems and solvers; in other words, the score-
S and P are the sets of solvers and problems respectively,
matrices of the problems and the solvers. Having these and J : S P R is an assessment function (a
matrices at our disposal, different rating/ranking performance or evaluation metric). It must be noted that
methods can be used. We illustrate our approach using each benchmarking context S , P, J , corresponds to
rating/ranking methods originated for ranking sports special multi-objective decision making (MODM) problem.
teams. The proposed approach is illustrated using an
example to demonstrate its viability and suitability for In this paper, we will offer new methods of
applications. benchmarking which are based on the special construction
of the score-matrices for problems and solvers in a
Keywords:- Benchmarking; Software; Solvers; Problems; benchmarking context S , P, J . Having these matrices
Testing; Ranking Method. at our disposal, well-known rating/ranking methods can be
used. Of course, rating/ranking methods can be very
I. INTRODUCTION different from one another. We illustrate our approach using
rating/ranking methods originated for the rating/ranking of
In recent years, intensive studies have been conducted sports teams.
to evaluate the effectiveness of various solvers. Several
methods for this purpose have been proposed in the The remainder of this paper is organized as follows:
literature review [1-9]. As noted in [9], most benchmarking section 2 describes the proposed methodology, section 3
tests utilize evaluation tables that display the performance considers the applications of the proposed tool in a selected
of each solver for each problem under a specific evaluation benchmarking problem, and section 4 contains a
metric (for e.g., CPU time, number of function evaluations, conclusion.
number of iterations, etc.). Different methods (based on
suitable “statistical” quantities) are used to interpret data II. METHOD
from these tables. The selection of a benchmarking method
currently depends on the subjective tastes and individual Throughout the paper, R n is an n-dimensional
preferences of researchers, who perform evaluations using
solver/problem sets and evaluation metrics. The advantages space, with norm || || and scalar product , .
and disadvantages of each proposed method are often a Moreover, we use the following notations for the special
source of disagreement; however, this only stimulates sets:
further investigation in the field. R n R n | k 0, k 1, , n ;
n
In our previous work [10], which was inspired by the n int R n | k 1 .
work [11], based on the concept of Revealed Comparative k 1
Advantage (RSA) [12] and the simplest version of the A. Benchmarking Context
PageRank method [13] [14], we proposed a benchmarking Consider a set P of problems and a set of solvers
method which was also tested on the particular example for S under the assumption that a function J : S P R ,
problems/solvers benchmarking. In this paper, we continue
henceforth referred to as the assessment function
our investigation on problems and solvers benchmarking
(performance metric), is given. Further assume, for
and propose new methods based on the special construction
definiteness, that the high and low values of J
of paired comparison matrices (score-matrices) of the
problems and the solvers. Having these matrices at our correspond to the “worst” and “best” cases respectively, and
disposal, different rating/ranking methods can be used. We for convenience interpret J ( s, p ) as the cost of solving
illustrate our approach using rating/ranking methods the problem p P with solver s S . Note that if
originated for sports teams ranking. J ( s, p ) J ( s, p), it can be said that s S solves
The approach which was introduced in [10] explores p P better than solver s S solves problem p P
the natural relations between problems and solvers, which is
B. Relation with MODM Problem Now we will offer the score-matrix S A . for
Let us assume now that S , P, J is a given alternatives. To emphasize the "sporting context" of this
consideration, let’s also note that we can interpret elements
benchmarking context. Define the set of alternatives as
of as athletes (e.g. chess players) who play the matches
1
def
A S s , , sn . At the same time, define the
S
among themselves and for each pair of athletes,
a, a A, their joint match (a, a) include m
criteria set C
c1, , cnP as follows:
games.
c j () J (, p j ) : S R ,
p j P p1 , , pnP , Now, for any a, a A, we define:
for j 1, , nP (i.e., we can assume that C P ). S A (a, a) scA (a, a),
cC
where
Hence, we obtain an MODM problem as a pair
1, c(a) c(a);
A, C where A is a set of alternatives and C is a scA (a, a) c C.
set of criteria. Note also that in the case under 0, c(a) c(a);
consideration, we can consider criteria as normalized and as
nonbeneficial (in the sense that low values correspond to Thus, the equality scA (a, a) 1 means that
the “worst” and the high values correspond to “best” case).
c(a) c(a) for criterion c C and the alternative a
Let’s recall now the basic concepts of multi-objective ("athlete a ") receives one score (i.e. "the athlete a wins
optimization theory [15, 16]. For a given MODM problem a game c C in the match (a, a) ). Correspondingly,
A, C where A a1 , , am is set of alternatives S A (a, a) indicates number of total wins of the athlete
and C c1 , , cn is a set of criteria, let’s introduce the a in the match (a, a). Obviously,
following notations and definitions: we say that A is the set m S A (a, a) 0, S A (a, a) 0, a, a A.
of admissible alternatives and map
c (c1 , , cn ) : A R n
is the criteria map We say that an alternative a has defeated an
(correspondingly c ( A) R n is the set of admissible alternative a, if S A (a, a) S A (a, a). We also say
values of criteria). Alternative a* A is weakly Pareto that the result of the match ( a, a). is " S A (a, a)
optimal (weakly efficient) if there is no a A with wins of the alternative a (losses of alternative a ),
c j (a ) c j (a* ) for all j 1, , n. Alternative a* A
score-matrix (of alternatives) for MODM problem and each ri is a measure of the performance of player
A, C .
i,1 i N , in the ranking problem N N , S . We
For any MODM problem A, C with the set of say also that for a ranking problem N N , S . a
alternatives A a1 , , am and the set of criteria ranking method R( N , S ) is induced by a rating vector
C c1 , , cn we can consider also adjoining MODM 1
r R N if :
problem C , A . Hence, we can define score-matrix for (i, j ) R( N , S ) if end only if ri rj .
the MODM problem C , A , which will be denoted as
C In this article, for illustrative purposes, we consider
S C S C (c, c) . Matrix S will be named as only a few ranking methods from the many that are known
c ,c C
score-matrix (of criteria) for MODM problem A, C . in literature2. In particular, for a score-matrix
S [ Sij ], 1 i, j N ,
D. Ranking Methods
Now we will briefly introduce the basic concepts of we consider the following ranking methods:-
ranking theory (see, for details, [17, 18] and the literature
cited therein). For a natural N , we say that a N N - Score Method
S [ Sij ], 1 i, j N , is a score-matrix if Rating vector r (r1 , , rN ) R , for the score
s s s N
matrix
method, defined as an average score
Sij 0, Sii 0, 1 i, j N .
N
To emphasize the "sporting context", lets note that ri s Sij / gi (S ),1 i N ;
we can interpret elements of N N 1, , N as j 1
2
Note that the ranking methods considered here originated from the
1
Note that when interpreting the results for the adjoined MODM ranking problems for chess tournaments and go back to the investigations
problem, we must remember the accepted agreement regarding of H.Neustadtl, E.Zermelo, and B.Buckholdz. For a detailed explanation,
"worst"/"best" values of criteria. If this causes inconvenience, then it is see e.g. [18].
better to carry out the appropriate normalization