Review Article | | Peer-Reviewed

A Systematic Review, Comparison and Juxtaposition of the (Dual Fitting and Factor Revealing LP Technique), (the ESMVERE CPLEX Algorithms) and the (Data Envelopment Analysis and TOPSIS Technique)

Received: 4 June 2025     Accepted: 17 June 2025     Published: 1 August 2025
Views:       Downloads:
Abstract

The objective of the present paper is to compare three location science methods. Two of the solutions discussed are practical real-world case studies, while the other method is used to reveal the approximation factor of an algorithm. We conduct an in-depth analysis of the originality, similarities, and differences of the three methods to unravel clarity on the conditions where each method may be applied, or where one method might be preferred over another. (1) The Data Envelopment Analysis (DEA)-Technique for Order Performance (preference) by Similarity to Ideal Solution (TOPSIS) technique is usable in many other areas of management science besides location analysis, however it is used in this case to guide the decision of selecting the best locations among alternatives. Hierarchical groups DEA super-efficiency and group TOPSIS technique is applied on mobile money agents’ locations in Harare Zimbabwe. (2) The ESMVERE algorithms introduce novel reformulations of the Uncapacitated Facility Location (UFL) and k-median problems which replace traditional distance matrices with Global Position System (GPS) based Euclidean Distance calculations. They are platform dependent solutions based on the CPLEX Optimization programming language (Opl). The solutions are used to solve a real-world problem which involve the near optimal siting of Hazardous Waste, Used Lead Acid Battery (ULAB), collection facilities in the Republic of Mauritius. (3) The Dual fitting with factor-revealing Linear Program (LP) technique is a location science approximation algorithm solution, used to analyze greedy algorithms that select facilities and connect the clients based on the dual solution information. The thrust of this review is to analyze the complexity, compare, contrast, and discuss the three location science techniques so as to classify them based on their purpose and area of application. Hence this review analyzes three papers and summarizes the main contributions of the three papers.

Published in American Journal of Operations Management and Information Systems (Volume 10, Issue 1)
DOI 10.11648/j.ajomis.20251001.12
Page(s) 13-37
Creative Commons

This is an Open Access article, distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution and reproduction in any medium or format, provided the original work is properly cited.

Copyright

Copyright © The Author(s), 2025. Published by Science Publishing Group

Keywords

DEA, TOPSIS, ESMVERE CPLEX UFL Problem Solver Algorithm, Dual Fitting, Factor Revealing LP

1. Introduction
The choice of the location of mobile banking facilities, Muvingi, Peer or mobile fast-food outlets which involve the “preference” of the decision maker, against the choice of the location of mobile clinics needed to administer vaccines or medication during a pandemic, also against the choice of the selection of locations of facilities needed to control and safeguard the whole population of a country against contamination from the hazardous waste of lead poisoning Mvere is the main contrast of this discussion . These real-world problems and solutions are contrasted against theoretical computer science techniques that guide facility location algorithm development, Jain, Mahdian and Mahdian, Spielman to determine their relationship, and to see why it may be extremely complex to apply some of the approximation algorithm techniques such as the Dual Fitting with approximation factor revealing LP technique to real-world problems . This technique is used to reveal the approximation factor of an algorithm.
DEA measures relative performance of a set of producers or Decision-Making Units (DMUs) where the presence of multiple inputs and outputs makes comparisons difficult . Hierarchical groups DEA super-efficiency (SE) and group TOPSIS technique is applied on mobile money agents’ locations in Harare Zimbabwe . It is used to make an optimal location analysis decision that incorporates multiple criteria such as, distance from the Central Business District (CBD), road density, zoning tariff, number of shops, number of mobile phone users, mobile transfer service users and population. It is a complex and time-consuming option, because it involves a lot of data gathering to process the decisions efficiently. However, to a greater extent, out of the three main location science techniques being reviewed in this paper it is the least complex in terms of the complexity, or level of hardness of the calculations performed.
DEA in general is an LP based technique solved in polynomial (P) time for a single model . TOPSIS is a deterministic, ranking -based method that also runs in P time . However, in more advanced formulations such as the Hierarchical groups DEA, it “often” becomes, non-deterministic polynomial time hard (NP-hard). The UFL problem and the k-median problem are already proved NP-hard, which implies that if P≠NP there is no algorithm known, which can find an optimal solution in P time . The level of hardness of the k-median problem is 1+2/e, proved by Jain, Mahdian . The level of hardness of the UFL problem is 1.463, proven by Guha, Khuller , it was proven NP-hard even in metric spaces where triangle inequality holds.
Group TOPSIS remains polynomial in standard form but can approach NP-hardness in extended frameworks . The formulas for Muvingi, Peer are quite easy to understand, they merely involve an understanding of what’s required at each stage, and an understanding of what each symbol represents so as to substitute them carefully and effectively with the data for calculation Figures 8, 9 and 10. These calculations compared to the lemmas of Jain, Mahdian and Mahdian, Spielman really bring out the differences in complexity, such as the complex calculations on Figures 1 and 2. The solution in Muvingi, Peer involves a lot of data input, thus it may not be as user friendly as the solutions in Mvere which simply involve the input of 3 types of data from the UFL problem which are GPS location coordinates of potential facility locations and client locations and the opening costs of the facilities. The solution in Muvingi, Peer may also again not be as user friendly as the solutions in Mvere which involve the input of 2 types of data from the k-median problem, which are GPS location coordinates of potential facility locations and client locations of the facilities. Thus Mvere solves a more complex problem with less input arriving at a calculated solution faster.
However, to a greater extent, in many cases that involve marketing and business value, the DEA and TOPSIS method could be a better choice and closer to a more accurate solution over all others, because it involves all other non-financial criteria that may be involved in the selection of a facility location. The choice of the preferable facility location depends on multiple criteria such as social, economic, environmental, demographic and many other subjective factors that are quantified to objective data in the decision-making process. This results in an all-encompassing decision which ranks all the alternatives or potential location sites based on the evaluation of a set of chosen criteria, resulting in a ranked list of all the potential sites.
The Dual fitting and factor revealing LP technique and the ESMVERE algorithms are primarily focused on minimizing the distances on an instance plane of just clients and facilities. The decision for the best choice of location is based on the calculation of the UFL problem or the k-median problem, both of which aim to cover a set of client locations, with a set of optimal facility locations, calculated from a set of potential facility locations. The main difference between the two being that the UFL problem also involves the facility opening costs while the k-median problem does not involve the facility opening costs, it just covers a set of clients with a set of open facilities, Figure 14.
The two problems answer the question of which set of facility locations should be opened from the larger set of potential facility locations, to ensure that all the clients receive service from an open facility near them at a minimum cost throughout the plane, for all the clients in an instance. ESMVERE algorithms go a step further by using GPS coordinate input for real world solutions. They also embed the Euclidean distance formulae in CPLEX Optimization Programming Language (Opl) as a single unified algorithm, which contrasts the traditional distance matrices used as input to solve these problems for the past 60 years .
Comparing the three methods in terms of originality, all three are based on original conceptualization . However, while Muvingi, Peer credibly extends the existing DEA and TOPSIS formulae, Jain, Mahdian and Mvere are developed on entirely new and original solutions to the UFL and k-median problems, that were clearly not in existence before they developed them, this could be more challenging.
“Some researchers have written custom software that directly structures and solves location models , relying on general optimization solvers. While custom software allows developers to specify their models, it comes with obvious shortcomings. It is challenging to develop new software, which requires specific technical skills—software design, tackling data input, model implementation and integration with solvers. ”.
Dual fitting involves constructing a feasible dual solution that is close to the optimal dual solution of a relaxed version of the original primal problem, using a relaxation of the constraints. This feasible dual solution is then used as a certificate, to guide the construction of a solution to the original primal problem . The methods solve a combinatorial problem which is complex to solve efficiently using the rule of thumb. They are focused on eliminating outliers and achieving the median, optimum distribution for all service centers of clients on a bipartite graph.
With DEA and TOPSIS the starting point is the choice of a set of potential sites (alternatives Chen, Li ), of a given number. The quest is then to sort the potential sites and rank them according to a set of criteria from best to worst. Muvingi, Peer uses 16 district and 49 suburb locations in the city of Harare Zimbabwe. This method is clerical and involves a lot of data gathering which may limit its optimum efficiency to only small size problems. Gathering efficient criterion data for, for example a 100 alternative locations appear exhausting, clerical and time consuming .
The choice of using district locations in Muvingi, Peer , is similar to the instance creation process of the ESMVERE algorithms, Mvere where one must choose what constitutes a potential facility location or a client location. Mvere use 156 gas stations GPS coordinates, as potential facility locations, while the client location coordinates are represented by the centroids of 144 districts locations. The two sets of GPS coordinate data cover the whole island country of Mauritius . These are the only sets of data required which are easier to gather efficiently. The opening cost of a potential facility location in Mvere , which is a main required input for the calculation of the UFL problem data, was collected from the potential facility location owners. The gas station owners decided how much they would charge to place a ULAB collection facility on their property.
In addition to the neat final solution quality Mvere , the ESMVERE CPLEX UFL algorithm showed superior computational efficiency as problem scale grew. Comparative analyses with other leading methods, including the Iterated Local Search (ILS), Multi-Start Descending Algorithm with Screening (MDAS), and hybrid heuristics (HYB), revealed that the ESMVERE CPLEX UFL problem solver algorithm consistently required lower average CPU time across all examined problem instances . This efficiency gain is particularly important in large-scale applications, where computational resources and time constraints play a critical role.
The Dual fitting and factor revealing LP technique was tested using randomly generated instances which vary in sizes from (20 facilities to 50 clients) to (150 facilities to 400 clients . These instances were generated on a 10000 x 10000 grid. The Internet Topology generator software, GT-ITM and the OR library were also used to generate test instances . The OR-Library is a library of test data sets for several operations research problems. Using this method on a real-world problem would mean at least the construction of a distance matrix of 20 facilities and 50 clients . That would mean 20 times 50 which is equal to 1000 pairs of distances in a distance matrix. It would mean measuring 1000 distances in the real world between potential facility locations and clients before the problem can be solved efficiently with this technique. For 156 potential facility locations and 144 clients’ locations in Mvere , it would mean measuring 22464 distances accurately to solve for a 1.861 approximate solution, Jain, Mahdian this would be a tedious data gathering process.
The algorithms in Mvere , are exact algorithms. An exact algorithm is a computational method that solves an optimization problem to its optimal solution, without any approximation or error . Exact algorithms can be computationally expensive, impractical and time consuming for large-scale problems because they aim to find the optimal solution . However, the inclusion of the Euclidean distance formulae to the UFL and k-median solution in Mvere has reduced computational complexity and made it usable even in large cases. “Sometimes, researchers develop specialized exact algorithms that exploit specific characteristics of the problem or its instances to achieve better performance than a general-purpose algorithm ”. These specialized exact algorithms made it possible for Mvere to solve the NP-Hard, UFL and k-median problems efficiently in small instances.
The algorithms in Mvere explore the entire solution space, which is extremely efficient in small cases of 7 facilities and 125 clients . However, for larger cases there is a need to purchase better versions of the software . The limitation on size meant that the island of Mauritius was divided into 9 regions and the selection of the best potential facility locations was calculated in series using the same client locations of each region with different facility locations. This calculation would select the optimum facilities to open in each region. This gave a near optimum result in Mvere compared to the optimum solution where the calculation was performed at once using the better version of the software . In most cases the same median facilities would be opened because the core concept remains the same, that each client connects to a facility which is near it. The difference was observed, with the optimal solution which performs the calculation at once for the whole island, that some regional borderline clients would connect to closer facilities in another region. Thus, the optimum solution had a lower value which went further in lowering the costs of the solution .
The approximation factor revealing LP technique on the other hand encodes the problem of finding the worst possible instance of a UFL problem as a linear program . This provides an LP family, one for each value of the number of cities or clients. Jain, Mahdian explains that the best value for the approximation factor (γ) is then the supremum (or the least upper bound) of the optimal solutions to these worst-case LP’s. Jain, Mahdian were unable; however, to explicitly measure the supremum, they get a solution which is feasible, to the dual of each of the worst-case LP’s. They conjecture that it is possible to determine an upper bound on the duals’ objective function value this is also an upper bound on the optimal γ. In the first algorithm, this upper bound is 1.861 .
There are situations where exact algorithms cannot be used these are cases that require Approximation algorithms . Approximation algorithms are used instead of exact algorithms when an exact solution is computationally infeasible, too slow, or when a near-optimal solution is sufficient . When finding an optimal solution is intractable, approximation algorithms offer a trade-off between speed and accuracy, finding solutions within a certain approximation factor of the optimal solution . The most common approach to overcome this difficulty, known as approximation algorithms, is by relaxing the requirement of finding optimal solutions . Instead, one can only hope to efficiently find a good approximate solution, one that is not optimal, but is within a small factor away from optimal . Many of these problems in optimization turn out to be computationally intractable, so a rational strategy is to settle for approximation algorithms, that is, algorithms that run in polynomial time and always have known near-optimal workable solutions .
2. Main Contents
2.1. Greedy Facility Location Algorithms Analyzed Using Dual Fitting with Factor-revealing LP
We analyze the first algorithm of Jain, Mahdian and Mahdian, Spielman which has an approximation factor of 1.861, to attempt to solve a real-world case of sitting ULAB collection facilities Mvere .
Jain, Vazirani proposed a 3 and 6 approximation algorithm for the metric UFL problem and the metric k-median problem, respectively. Their algorithm is an extension of the primary-dual scheme to deal with a primary-dual pair of LPs that are not a covering-packing pair for the UFL metric problem and the k-median metric problem.
To this day, the algorithm they suggest forms the key core basis for all of the primal dual methods. The basic idea behind the algorithm is that each city j continues to boost its dual variable, αj, until it is linked to an open facility, other primary and dual variables simply respond to this transition, trying to preserve feasibility or satisfy complementary conditions of slackness .
Jain and Mahdian formalize the methodology used as the dual fitting technique, to examine the set cover problem, Mahdian, Spielman , also introduce the concept of using the approximation factor-revealing LP. Using this duo, Jain and Mahdian analyze two greedy algorithms for the metric UFL problem with 1.861 and 1.61 approximation factors for both problems, respectively, and O(mlogm) and O(n3) runtimes, respectively. When explaining the dual fitting method they imply that the basic algorithm is combinatorial, and is a simple greedy algorithm in the case of the set covering problem, assuming it to be a problem of minimization.
Jain and Mahdian describe the combinatorial algorithm as a primal-dual-type algorithm that allows primal and dual updates iteratively, using the problem’s linear programming relaxation and its dual-type algorithm. Although it is not a primal dual algorithm, since the computed dual solution is infeasible, they show that the primal integral solution found by the algorithm is fully paid for by the dual computed solution, meaning that the primal solution’s objective function value is bounded by that of the dual solution . The algorithm’s main step is to divide the dual by a suitable factor, γ, and to prove that the shrunk dual is feasible, i.e. fits into the instance provided.
On OPT, the shrunk dual is then a lower limit, and γ is the approximation guarantee of the algorithm. The aim is therefore to find the necessary minimum γ, which is to find the worst possible instance of the problem, one in which the dual solution needs to be shrunk as much as possible in order to make it feasible . Jain and Mahdian describe an approximation factor-revealing LP that encodes the problem of finding the worst possible instance as a linear program.
This provides an LP family, one for each value of the number of cities nc. Jain and Mahdian further explain that the best value for γ is then the supremum of the optimal solutions to these LP’s. They were unable however, to explicitly measure the supremum, they get a solution which is feasible, to the dual of each of the LP’s. They conjecture that it is possible to determine an upper bound on the duals’ objective function value this is also an upper bound on the optimal γ. In the first algorithm, this upper bound is 1.861 and in the second algorithm, 1.61. Jain and Mahdian numerically solves the factor-revealing LP with a large value of nc to get a closely matched tight example.
The first algorithm of Jain and Mahdian , is similar to the greedy set cover algorithm, which iteratively selects the most cost-effective choice at each stage. Cost effectiveness is defined as a measure of the ratio of the costs incurred to the number of new cities served. To evaluate this algorithm, they used LP-duality and gave an alternate definition that can be seen as a Jain, Vazirani algorithm update.
Algorithm 1 enhances facility location efficiency by balancing opening and connection costs through a structured cost comparison and iterative refinement process. This approach leads to optimized facility placement and city assignment, reducing total operational costs.
The most basic explanation of Jain and Mahdian 's first algorithm is, when a city is connected to an open facility, it withdraws everything it has contributed to the opening costs of other facilities, meaning that the dual pays for the primal solution in full.
Jain and Mahdian summarise their first algorithm as:
Let U be an unconnected cities set.
All cities are unconnected at the beginning, i.e. U:= C, and all facilities are unopened.
IfU  
Of all the stars, find the most cost-effective one (i, C'), open i if it’s not already open, and connect all the cities in C' to i
fi:=0,U:=U \ C'
A star consisting of one facility and several cities is defined by Jain and Mahdian . The cost of a star is the sum of the facility’s opening cost and the connection costs between the facility and all the cities in the star in their algorithm. This is formally clarified as the price of a star (i,C'), where i is a facility, and C'  C is a subset of cities, fi + (jC')cij. A facility can be selected again after opening in their algorithm, but the opening cost is counted only once and fi is set to 0 after the first algorithm selects the facility. When connected to an open facility, each city j is deleted from C and is not taken into account again.
An order is carefully observed with the two algorithms for each facility i, they sort the cities in increasing order of their connection cost to i. Jain and Mahdian determine that a facility and a set containing the first k cities in this order for some k will consist of the most cost-effective star. They describe αj as the city j’s contribution or its share of the overall total expenses. They explain how raising the dual variables in each iteration of the greedy algorithm will help find the most cost-effective star.
They restate Algorithm 1 based on the observation that, the most cost-effective star will be the first star (i, C') for which (jC')max(0,αj-cij)=fi. Dual variables of all unconnected cities are simultaneously increased.
Jain and Mahdian Algorithm 1 restatement:
They add a notion of time, such that each occurrence is related to the time it occurred. The algorithm begins at the moment 0. Each city is initially set to be unconnected (U:= C), all facilities are unopened, and αj is set to 0 for each j.
While U   increases the time and, simultaneously, j  U increases the αj parameter at the same rate for each city, until one of the incidents below takes place (if two events occur at the same time, they are processed in an arbitrary order).
For some j of an unconnected city, and for some i of an open facility, αj=cij. Connect city j to facility i and delete j from U in this case.
We’ve got (jU)max(0,αj-cij)=fi for some unopened i facility. This ensures that the cities’ total contribution is adequate to open the i facility. Open this facility in this case, and j with αj  cij for each of the unconnected cities, connect j with i, and delete it from U.
Jain and Mahdian give an analysis of algorithm 1 in which they explain that every city’s contribution goes into opening one facility at MOST and connecting the city to an open facility. They determine that the total cost of the solution generated by their algorithm is equal to the amount of the jαj contributions. Jain and Mahdian explain that α is not a feasible dual solution since they remove a subset of cities in each iteration of the restatement of algorithm 1 and withdraw their input from all facilities.
At the end of a facility, therefore, i, jmax(αj-cij,0), can be greater than fi, thereby violating the corresponding dual constraints. They reiterate that the quest is to find a suitable γ for which α / γ is feasible, jαj/γ will be the lower limit to the optimum, so at most γ will be the approximation factor for the algorithm . They include a description that describes that if α / γ is a feasible dual, it is trivial if and only if each facility is γ-overtight at most .
jmax(αj/γ-cij, ,0) fi(1)
They conjecture that the goal is to find such a γ, and thus the need is to only consider the cities j for which αj  γcij. Jain and Mahdian give a lemma which captures the metric property:For every two cities j,j' and facility i, αj  αj' + cij' + cij. This lemma is illustrated in the worked example diagrams City A Algorithm 1 and City B Algorithm 1. They state that when αj'  αj, the inequality obviously holds, as on the diagram City B Algorithm 1.
However as on the diagram City A Algorithm 1 assuming αj > αj'. If city j' is connected to facility i' as on the diagram City A Algorithm 1, city A is connected to facility X, fi=10, this facility i' which is fi=10 on the diagram is open at time αj' which is at 5 on the diagram. The contribution αj which is 8 on the diagram cannot be greater than ci'j which is the distance from client C to facility X, fi=10 on diagram, because in that case city j or city C could be connected to facility X, i' or facility fi= 10 at some time t= 5 < αj=8.
Figure 1. City A Algorithm 1.
Figure 2. City B Algorithm 1.
The inequality calculations on Figure 1 serve as a validation step within the algorithm to determine whether a particular city to facility assignment is feasible or optimal based on cost constraints. The inequality compares the sum of connection costs and parameters for two facilities i and i’ with respect to a city j. it checks if the combined cost of assigning city j to facility i’ plus the facility opening cost fi’ is less than or equal to the cost of assigning the same city to facility i plus fi. This helps decide if swtching the city’s assignment to a different facility reduces the overall cost.
The result “UNTRUE” indicates that the assignment is not cost-effective to change under the given parameters. This step is crucial for the algorithm to iteratively refine facility openings and city assignments to minimize total costs in the location optimization problem.
Algoritm 1 improves facility location efficiency by systematically evaluating and optimizing the assignment of cities to facilities based on combined costs, including fixed facility opening costs and connection costs. It calculates the total cost for assigning cities j and j’ to a facility i using the fomular αjαj'+cij'+cij where αj represents contributions of each city towards opening a facility and cij represents the connection costs. By comparing these costs across different facilities and cities, the algorithm identifies the most cost-effective assignments, ensuring that facilities with lower total costs are prioritised and opened, while less efficient ones are removed from from the consideration set
The algorithm iteratively opens facilities with the lowest combined costs and removes those that do not improve the overall cost structure, as shown by the removal of facilities W, X and Y from the set U. This selective opening and closing process reduces unnecessary facility openings, minimizing fixed costs while maintaining service quality. The cost comparisons, such as ci'j+cij'+cijαj'+cij'+cij ensure that city assignments are only changed when it leads to a net cost reduction, thus improving allocation efficiency.
Table 1 contains the set of all the cities on the map instance. The map has 16 cities on the first column to the left, On the second column each city is labeled after the facility it is connected to, its nearest facility, either facility W, facility Y or facility X. The city length is the distance between the city and its nearest facility. After a facility opens the city connects to the facility, this is the time it is removed from U, the set of connected facilities.
Table 1. Restatement of algorithm 1.

U set of cities from Figure 1

city Label

city length

The time its removed from U

1

W1

1

8

2

W2

2 city A

5

3

W3

2

8

4

W4

3

8

5

W5

3

8

6

W6

4

8

7

W7 city B

4

8

8

Y1

1

12

9

Y2

1

12

10

Y3

2 city B

8

11

Y4

5

12

12

Y5

5

11

12

X1

1

5

13

X2

1

5

14

X3

3

5

15

X4

3

5

16

X5 city A

3

5

Table 2. Facility W fi=30.

t=αj

W1

W2

W3

W4

W5

W6

W7

fi contributions

dj

cij

1

2

2

3

3

4

4

=19

αj

time

αj1

αj2

αj3

αj4

αj5

αj6

αj7

Contributions

0

0

0

0

0

0

0

0

0

1

1

1

1

1

1

1

1

0

1

2

1

1

1

1

1

1

1

1

4

3

1

1

1

1

1

1

1

3

9

4

1

1

1

1

1

1

1

5

15

5

1

0removed

1

1

1

1

1

6

21

6

1

0

1

1

1

1

1

6

27

7

1

0

1

1

1

1

1

6

open 30

@8

1

0

1

1

-

-

-

3

αjs

αjs

8

4

8

8

7

7

7

Table 2 shows the contribution of each city in terms of αjs towards the opening of the Facility W. The αj increases as time in seconds, while each connected city contributes towards the opening cost off the facility W. Each city contributes towards covering its distance its distance (cm) first, bold font numbers, when it reaches the facility in distance (cm), then it starts contributing towards the opening of the facility (30 units). When all the contributions from all the cities connected to the facility are equivalent to the opening cost of the facility, the facility opens and connects every city at time 8.
City W2 is connected to another facility from the start of the time facility X, hence when the other facility X opens it connects to it. At time five its removed from U hence it stops contributing towards facility W opening
Table 3. Facility X fi=10.

t=αj

X1

X2

X3

X4

X5

fi contributions

dj

cij

1

1

3

3

3

=11

αj

time

αj1

αj2

αj3

αj4

αj5

Contributions

0

0

0

0

0

0

0

1

1

1

1

1

1

0

2

2

1

1

1

1

1

2

4

3

1

1

1

1

1

2

9

4

1

1

1

1

1

5

Open 10

@5

1

-

-

-

-

1

αjs

αjs

5

4

4

4

4

Table 3 shows the contribution of each city in terms of αjs towards the opening of the facility X. The αj increases as time in seconds, while each connected city contributes towards the opening cost of the Facility X. Each city contributes towards covering its distance (cm) first, bold font, when it reaches the facility in distance (cm), then it starts contributing towards the opening of the facility (10 units). When all the contributions from all the cities connected to the facility are equivalent to the opening cost of the facility, the facility opens and connects every city at time 5.
Table 4. Facility Y fi=40.

t=αj

Y1

Y2

Y3

Y4

Y5

fi contributions

dj

cij

1

1

2

5

5

=14

αj

time

αj1

αj2

αj3

αj4

αj5

Contributions

0

0

0

0

0

0

0

0

1

1

1

1

1

1

0

2

2

1

1

1

1

1

2

5

3

1

1

1

1

1

3

8

4

1

1

1

1

1

3

11

5

1

1

1

1

1

3

16

6

1

1

1

1

1

5

21

7

1

1

1

1

1

5

25

8

1

1

0removed

1

1

4

29

9

1

1

0

1

1

4

33

10

1

1

0

1

1

4

37

11

1

1

0

1

1

4

Open 40

@12

1

1

0

1

-

3

αjs

αjs

12

12

7

12

11

Table 4 shows the contributions of each city in terms of αjs towards the opening of the Facility Y. The αj increases as time in seconds, while each connected city contributes towards the opening cost of the Facility Y. Each city contributes towards covering its distance (cm) first, bold font, when it reaches the facility in distance (cm), then it starts contributing towards the opening of the facility (40 units). When all the contributions from all the cities connected to the facility are equivalent to the opening cost of the facility, the facility opens and connects every city at time 12.
Jain and Mahdian give another lemma which shows how the sum of the αjs of every facility must always be below the opening cost. They state this as, during the algorithm, the total contribution provided to a facility at any time is no more than its cost. For every city j and facility i l=jkmax(αj-cil) fi. The tables W, X, Y of the worked example diagram Algorithm 1 city B are an example of this. For example Facility W, j=W1W7max(49-19) 40, For Facility X, j=X1X5max(21-11) 10 and For Facility Y, j=Y1Y5max(54-14) 40.
In this lemma Jain and Mahdian give an example of a αk  αj for which the αk algorithm stops increasing as soon as the k city has a tight edge to a facility which is fully paid for. This means that when a facility opens the algorithm stops increasing the αjs thus there is with many facilities at least one αk which does not grow to the size of the αj. Facility X is an example of this, with city X1 being the αj @5 and cities X2, X3, X4, X5 being the αks which stop growing at 4.
Jain and Mahdian state that the goal is to find the minimum γ for which j=1k(αj/γ-cij) fi, hence the maximum of the ratio j=1kαjf+j=1kdj is sought.
They go on to state that zk the maximum value of the objective function of the following program (2) is equal to the Linear Program (7) which they call the factor revealing LP below.
zk=maximizej=1kαjf+j=1kdj(2)
Subject to the constraints:
αjαj+1 j{1,..........,k-1}(3)
To note this is explained by Facility X where there is a k which can be the city larger than all the other cities explained by city XI being the k and cities X2, X3, X4, X5 being the ascending αjs 1,.........., k-1.
αjαl+dj+dl j,l{1,..........,k}(4)
l=jkmax(αj-dl,0)j{1,..........,k}(5)
αj,dj,f0 j{1,..........,k}(6)
The Factor Revealing LP
zk=maximizej=1kαj(7)
Subject to the constraints:
f+j=1kdj1(8)
αjαj+1 j{1,..........,k-1}(9)
αjαl+dj+dl j,l{1,..........,k}(10)
xjlαj-dl j{1,..........,k}(11)
l=jkxjlj{1,..........,k}(12)
αj,dj,f0 j{1,..........,k}(13)
They state that (Primal):
1) zk is the optimal solution to the approximation factor-revealing LP.
2) zk is a lower bound on the value of γ.
3) zk is the minimum γ for which j=1k(αj/γ-cij) fi.
4) zk is the maximum value of the objective function of the ratio j=1kαjf+j=1kdj.
At optimal, primal and dual solutions are equal thus:
They state that (Dual):
1) A solution to the dual of the approximation actor-revealing LP is needed to prove an upper bound on γ.
2) An upper bound on the objective function values of the duals of the factor-revealing LP is an upper bound on γ.
3) It is possible to solve the dual of the factor-revealing LP to prove a close to optimal upper bound on the value of zk.
The suprenum being the least upper bound, The suprenum of (zk) the optimal solution to the approximatimation factor-revealing LP supk1{zk} is the best value of γ. Therefore γ =: supk1{zk}.
Jain and Mahdian give a lemma for every k  1, zk  1.861.
They multiply the inequalities of the factor-revealing LP to give an inequality of the form:
j=1kαj/γ-j=1kcijf(14)
j=1kαj-γj=1kcijγf(15)
j=1kαj-1.861j=1kdj1.861f(16)
They suggest that finding the right multipliers for the above inequalities is equivalent to obtaining a feasible solution to the dual of the factor-revealing LP. They add three constants p1, p2 and θj to prove the bound on γ for algorithm 1.
i=jkθj1.861γ(17)
j=1kαj==j=1p1ki=jp2kθj(αj-di)(18)
-1.861j=1kdj==j=p1k+1ki=jkθj(αj-di)(19)
i=jkθjf==1.861f(20)
coeffA[αj]=(p2k-j+1)θjif jp1k(k-j+1)θjif j>p1k(21)
coeffA[-dj]=i=1jθiif jp2ki=p1k+1jθiif j>p2k(22)
From equation (18), (19) and (20) the inequalities (16) and (28) are compared. Inequality (28) extends inequality (16) by adding the three constants p1, p2 and θj. They denote that the left hand side of inequality (28) by A and furthermore propose that the sum of the coefficients of αj’s in A is at least k. Thus θj’s are needed to satisfy the inequality:
j=1p1k(p2k-j+1)θj+j=p1k+1k(k-j+1)θjk(23)
They propose inequality (23) which guarantees that the average coefficient of αj’s in A is at least one. An example of this is the best way to describe the concept: If k is 20, the number of cities being 20.
j=120αj(24)
The total sum of αj’s could be 23 therefore
j=120αj=23(25)
Dividing 23 the total sum of αj’s by their number 20 then the average coefficient of αj’s is 1.5.
Jain and Mahdian propose that for: lj  j
i=jlj(αj-di)f(26)
lj=p2kifjp1kkifj>p1k(27)
j=1p1ki=jp2kθj(αj-di)+j=p1k+1ki=jkθj(αj-di)i=jkθjf(28)
They prove that:
1.861==γ(29)
in the worst case.
It should be possible, from the look of things by simple analysis, to use an instance created, for the reverse supply chain network design of ULABs in Mauritius, Mvere to achieve a solution guaranteed by γ. Since γ guarantees that the greedy algorithm Jain and Mahdian solves the UFL problem, (1.861) close to an optimal solution. However, this is not the purpose of this algorithm, and achieving its use in the real world would be extremly difficult and complex. An attempt at this would mean:
1) The instance created for siting ULABs in Mauritius Mvere has to be created.
2) To create this instance there is a need to develop a Distance matrix of 156 facilities and 144 clients
3) This n x m would be 22464 distances that have to be measured for each pair of distances between a client and a facility
4) The instance created for siting ULABs in Mauritius are encoded as a family of linear programs.
5) The same instance is recreated into a family of instances with different number of nc or cities and clients.
6) Solve these family of LPs created, to get their optimal solutions.
7) Obtain the suprenum of these LPs and compare it to the already known (γ).
8) In each case, with different number of cities nc, we have a usable optimal solution guranteed by (γ).
9) This is because (γ) is the supremum (or the least upper bound) of the optimal solutions to these worst case LP’s.
This has so far proved to be a misconception, an asumption that since the algorithm is well formulated in theory, it can be easily attempted to solve a real world case. Many theoratical science approximation algorithms such as Jain and Mahdian have the most complex transfer from science research to real world application. Approximation algorithms in general can be adapted to specific needs and constraints, they can be used to find multiple solutions, allowing for greater flexibility in decision-making, especially when dealing with real-world problems where constraints and objectives are often "approximately" formulated . However, not all approximation algorithms are usable practically in real world problems. Some involve complex data structures, or sophisticated algorithmic techniques, leading to difficult implementation issues or improved running time performance (over exact algorithms) only on impractically large inputs .
In consultation with the Author of , who is an emminent expert and scolar in the field of theoratical computer science, they revealed that they had not, at the time of writing, implemented their approximation algorithm of in real world practice, and they were not very familiar with the application side of the two problems, the UFL and the k-median problem. “If you want to implement algorithms for the two problems, you may try to start with the primal-dual algorithms since they do not require you to solve a linear programming” .
The primal-dual approach in optimization does not always require solving a linear program directly to find a solution . It's a powerful technique that constructs feasible primal and dual solutions iteratively, often leading to algorithms where analysis is derived from the primal-dual interaction . It focuses on iteratively improving feasible solutions and using the duality relationship to gain insights into the problem, reducing the duality gap, and eventually finding the optimal solution . Primal-dual algorithms, avoid explicitly solving a (LP) problem . Instead, they construct approximate solutions to the problem and feasible solutions to the dual of an LP relaxation simultaneously . This approach leverages the duality relationship between the primal and dual problems, allowing for a more efficient and insightful solution process .
2.2. The Near Optimal Siting of Hazardous Waste (Used Lead Acid Battery (ULAB)) Collection Facilities in the Republic of Mauritius Using the ESMVERE CPLEX UFL Problem Solver Algorithm, ESMVERE CPLEX k-median Problem Solver Algorithm
Mvere explores reverse supply chain network design problems in the automotive industry of the republic of Mauritius, specifically focusing on the challenge of recycling ULABs. The authors examine logistical issues related to the UFL and the k-median location problems . Thus Mvere deals with the important problem of improving the safety and effectiveness of the collection, processing, and disposal of one of the most hazardous parts of the vehicles, the ULAB .
Mvere presents a mathematical framework that consists of two facility location problems the UFL, and the k-median problems as alternative approaches to finding the best location sites for the collection facilities of the ULAB batteries Mvere present an innovative approach to the siting of hazardous waste collection facilities in Mauritius using the novel ESMVERE CPLEX UFL problem solver algorithm, and the ESMVERE CPLEX k-median problem solver algorithms . Both papers address a significant gap in the literature by applying UFL and k-median problem-solving techniques in a real-world setting, using the actual data from the Mauritius statistical office . Mvere ,illustrated a step-by-step instance creation process which familiarizes the reader with how to solve a realistic real-world problem. They clearly demonstrated the algorithm solution to the real-world data from Mauritius using the ArcGIS mapping . The solutions are also visible when they are input on google earth .
The primary contribution of Mvere is the removal of the distance matrix as input in the calculation for the UFL problem and the k-median problem, replacing the distance matrix with the GPS input and Euclidean distance formulae . This simplified the data input and reduced computational complexity for the algorithm making it easy, usable, and convenient as a solution .
Figure 3. UFL Matrix solution. Source. Cornuejols, Nemhauser
Since their introduction over 60 years ago by Balinsky and Hakimi both the UFL and k-median problems have been solved using the distance matrix. Jain, Mahdian , Cacioppi, Watson , Mistry , Mahdian, Spielman , are examples of the CPLEX Opl software algorithms that solve the UFL problem using the input of the distance matrix Srinivasan is a p-median example of how the problems were solved . Cornuejols, Nemhauser is an example of how the UFL problem was solved, as seen on Figure 3.
Thus Mvere introduce a solution that does not need the input of a distance matrix but uses GPS based Euclidean distance calculations, while, to the best of our knowledge, all known prior solutions used the input of a distance matrix in the calculation of the UFL problem and the k-median problem .
It would be incredibly difficult to use this solution in a real-world scenario, which could explain why there are not so many real-world examples . All known prior innovations to the UFL and k-median problem solutions were centered on improvements and innovations around the matrix . Thus Mvere proposes a solution which is a novel breakthrough spanning over 60 years of work in this area . Cacioppi, Watson , use the distance matrix below (30), in a small nine warehouse example . Even in their short example, to ensure that the distance of each point to the other remaining eight points is combinatorically considered, is a big combinatorial task . This becomes even more complex when you must gather the data between all possible combinations of distances, for example if there are 156 facility locations and 144 client locations m x n = 22464 location distances to measure and place in a distance matrix. This is somewhat easy when it is randomly generated by software, however, in real world practice this is extremely complex. Below is a distance matrix of only 9 x 9 locations .
Cacioppi, Watson Distance Matrix (30)
Distances [[0, 720, 790, 297, 283, 296, 461, 796, 996]
[720, 0, 884, 555, 722, 461, 685, 245, 1099]
[790, 884, 0, 976, 614, 667, 371, 645, 219]
[297, 555, 976, 0, 531, 359, 602, 715, 1217]
[283, 722, 614, 531, 0, 263, 286, 629, 721]
[296, 461, 667, 359, 263, 0, 288, 479, 907]
[461, 685, 371, 602, 286, 288, 0, 448, 589]
[769, 245, 645, 715, 629, 479, 448, 0, 867]
     [996, 1099, 219, 1217, 721, 907, 589, 867, 0]];
This would be impractical in many cases in the real world where there might be hundreds or thousands of location points, and the distances that make each pair of distances . Thus, a new model is developed that does not need the input of the distance matrix but uses the formulae for the Euclidean distance between two points and the input of actual GPS coordinate location points to solve the UFL and the k-median problem in Mvere . This solution is user friendly and convenient.
The formula for the Euclidean distance between two points is:
d=x1-x22+y1-y22 (31)
The UFL, k-median CPLEX Opl Euclidean distance formulae developed for this study by Mvere :
function getDistance ulabcity1, ulabfacilitycity1
{
return opl.sqrt(opl.pow(ulabcity1.x-ulabfacility1.x,2)+opl.pow(ulabcity1.y-ulabfacility1.y,2)) 
}
The new method solves the UFL and the k-median problem by making a choice between facility locations and client locations, without considering the distance between all the points as a distance matrix . The newly developed method only requires the GPS coordinate location points . This means that it opens the best out of n potential facility locations and assigns only m clients to these open facilities per calculation .
A combination of the CPLEX solutions to the Travelling Salesman Problem (TSP) by Caceres and a CPLEX solution to the p-median algorithm by Cacioppi, Watson , was used to develop a new and unique software invention which solves the UFL and k-median problem by Mvere [1 ,2 ,6]. The works of Cacioppi, Watson and Caceres were thoroughly analyzed and reviewed.
Combining these two solutions made it possible to develop 3 entirely new CPLEX models which solve the UFL problem, k-UFL problem and k-median problem . The sub tour elimination decision variables of the TSP problem were removed . Arrays of the facility locations and opening costs were added in Mvere , these enabled the costs of the edges to be computed from each facility to each client. The changes have developed entirely new software solutions in Mvere after careful consideration and study.
Definition of parameters
a) i is used to denote a (ulab collection facility).
b) j is used to denote a (ulab city) it is a residential area with a population of clients that want to dispose ulabs in Mauritius.
c) xij is a decision variable which denotes whether (ulab collection facility) i is connected to (ulab city) j or not. xij is a binary variable, which becomes 1 if (ulab collection facility) i is connected to (ulab city) j or 0 if (ulab collection facility) i is not connected. Being connected means the (ulab city) can dispose its ulabs at the open (ulab collection facility).
d) yi is a decision variable which denotes whether (ulab collection facility) i is open or not. yi is a binary variable, which becomes 1 if (ulab collection facility) i is open to receive disposed ulabs from (ulab cities), or 0 if (ulab collection facility) i is not open to receive ulabs.
e) cij denotes the cost of the distance from (ulab city) j to (ulab collection facility) i. The distance determines the cost which has to be minimized. The distance is only a significant cost if the (ulab collection facility) i is open and connected to (ulab city) j, that is if xij is 1.
f) F is the larger set of (ulab collection facilities). The goal is to open the best set of (ulab collection facilities), which serve all the (ulab cities) at a minimum cost. The best set of (ulab collection facilities) is chosen and opened from the larger set of (ulab collection facilities).
g) C is the larger set of (ulab cities). All the (ulab cities) have to be served and Connected to a (ulab collection facility). All the (ulab cities) have to be connected to an open (ulab collection facilities), Unlike and direct opposite of (ulab collection facilities, where there is a choice to select the optimum located, open them, and let the rest remain closed.
h) F' is the best set of (ulab collection facilities) opened. This is when the best set of (ulab collection facilities) is chosen and opened from the larger set of potential (ulab collection facilities).
i) fi the costs of opening (ulab collection facilities) i. The facility opening cost is one of the variables that determines the final value of the objective function. The objective is to minimize the sum of the (ulab collection facilities) opening costs.
The metric UFL problem is formulated as an integer program to satisfy the constraints:
MinimizeiFfiyi+(iF,jC)cijxij.(32)
Subject to the constraints:
(iF)xij=1 jC(33)
xijyi iF,jC(34)
xij0,1 iF,jC(35)
yi0,1 iF(36)
The mathematical formulation of the UFL problem is then converted to the CPLEX Opl language. The model formulation is translated into the Opl language using the IBM ILOG CPLEX Optimization Studio IDE 20.1.0. Exactly as the TSP model by Caceres described in the literature the CPLEX Optimization Engine requires a model of the form:
1. Size or Range
2. Input Data Sources
3. Decision Variables
4. Decision Expressions
5. Objective Function
6. Constraints
1) Size or Range
The Opl language is used to declare two integer variables n is for the (ulab collection facilities) and m is for the (ulab cities). The Opl language is used to create two range variables facilities and cities which contain all the (ulab collection facilities) and (ulab cities) on the plane.
int n=...; // is the number of (ulab collection facilities)
int m=...; // is the number of (ulab cities)
range ulabfacilities=1..n; //the range of all the facilities from 1 up to n.
range ulabcities=1..m; //the range of all the cities from 1 up to m.
int fi[ulabfacilities]=...; //this is an array of the opening costs of all (ulab collection facilities)
In the CPLEX data file //the opening costs of all (ulab collection facilities) are given as a matrix:
fi=[30, 20, 50, 20, 40, 60, 10];
2) Input Data Sources
The tuple Location is created for this model, this contains the float variables x and y. When it is called, this method divides the plane into x and y coordinates. It gives the location of every (ulab collection facility) and (ulab city). It also gives the starting and ending point of every edge, in terms of its x and y coordinates.
tuple Location {float x;float y;}
The tuple edge is created, this contains the integer variables i and j. When it is called, this method connects all the points on the plane into edges, and every edge has beginning point (ulab city) j and ending point (ulab collection facility) i. Thus, the tuple edge is called to create the edges.
tuple edge {int i;int j;}
A set of edges named edges is created, which defines the character of what an edge is, and defines its boundaries, while placing all the edges on the plane into a set. The edge is described as less than the variable i and greater than the variable j, i is found in the rangeulabfacilities=1...n; and j is within the rangeulabcities=1...m;.
setof (edge) edges={<i,>| i in ulabfacilities, j in ulabcities };
A float variable c is created, this is an array variable. c is an array of all the edges from the variable edges which contains the set of all the edges. Square braces [] are used to denote an array and in this case are used to describe the variable c as an array which contains the set of all edges edges.
float c[edges];
The tuple Location is called to create an array ulabcityLocation which contains the range ulabcities. This array places x and y coordinates in every city. The array stores the x and y coordinates of every city.
Location ulabcityLocation [ulabcities];
The tuple Location is called to create an array ulabfacilityLocation which contains the range ulabfacilities. This array places x and y coordinates on every facility. The array stores the x and y coordinates of every facility.
Location ulabfacilityLocation [ulabfacilities];
The generated data is preprocessed before the problem is computed. To preprocess the data, a code block execute{} is created. In the code block, a functiongetDistance() is declared, this function calculates and returns the Euclidean 2-dimensional distance between a facility and a city. It returns this value based on their x and y coordinates. The formula for the Euclidean distance between two points is the same as on formular (31)
d=x1-x22+y1-y22
calculated by subtracting the x coordinates of the two points and the result squared is added to the subtraction of the two y coordinates of the two points. The difference of the subtraction of the two points were the x coordinates are subtracting each other and the y coordinates are subtracting each other is squared and then it is added. The square root of this difference is the length of the distance. This general common formula is translated into the Opl Language in the function getDistance() illustrated in the code below.
In the code block execute{}, a for loop which randomly generates the x and y coordinates of every ulabcity are created as well as a for loop which randomly generates the x and y coordinates of every ulabfacility. Finally in the same code block execute{}, a for loop with a variable e is constructed. e represents every edge in the set of edges. The loop goes to the array c which is an array which stores all the edges from the variable edges which contains the set of all the edges, it then uses the functiongetDistance() to calculate the size of every edge in terms of the x and y coordinates of its starting point the ulabcityj and its ending point the ulabfacility i.
execute {
function getDistance (ulabcity1, ulabfacilitycity1) {
return opl.sqrt(opl.pow(ulabcity1.x-ulabfacility1.x,2)+opl.pow(ulabcity1.y
-ulabfacility1.y,2)) }
for(var i inulabfacilities) {
ulabfacilityLocation[i].x;
ulabfacilityLocation[i].y;
}
for(var j inulabcities) {
ulabcityLocation[j].x;
ulabcityLocation[j].y;
}
for(var e in edges){
c[e] =getDistance(facilityLocation[e.i],cityLocation[e.j])
}
The code block execute{} above, which is written in the Opl Language, contains one function and three for loops. The function getDistance() returns the Euclidean 2-dimensional distance between two points.
The first for loop in the code block execute{} is created by declaring a variable i. The variable i loops through every (ulab collection facility), in the range ulabfacilities. It places x and y coordinates. The code ulabfacilityLocation[i].x; searches for an x coordinate from the array ulabfacilityLocation in the data file.dat it loops through all the (ulab collection facilities) in the array. The code facilityLocation[i].y; searches for an y coordinate from the array ulabfacilityLocation looping through every (ulab collection facility) in the array.
The second for loop in the code block execute{} is created by declaring a variable j. The variable j loops through every (ulab city), in the range ulabcities. It places x and y coordinates. The code ulabcityLocation[j].x; searches for an x coordinate from the array ulabcityLocationin the data file.dat it loops through all the (ulabcities) in the array. The code ulabcityLocation[j].y; searches for an y coordinate from the array ulabcityLocation looping through every (ulab city) in the array.
The third for loop in the code block execute{} is created by declaring a variable e. The variable e loops through every edge in the set of edges. The function getDistance() then calculates the distance between the x and y coordinates of the i and the x and y coordinates of the j of each edge. Each edge has beginning point j and ending point i. The function getDistance() returns this distance for each edge in the array of edges c[e]. This just gives the length of every edge on the plane.
3) Decision Variables
There are two main decision variables in the UFL problem xij, which is declared in CPLEX Opl in the model as a Boolean decision variable
dvar boolean x[edge];
, then there is yi which is also declared in the model as a Boolean decision variable
dvar boolean y[facilities];
The two variables stand for the solution set of facilities and edges. The variable x is multiplied to an element of an array []. The array that the variable x is multiplied to, is an array of the tuple edge. The elements of the tuple edge are the contents of the array [edge]. Thus, the tuple edge contains integer variables i’s and j’s. Therefore x connects i’s and j’s in an array to create the edges.
4) Decision Expressions
Decision variables and decision expressions are used to create the actual Opl model. Caceres translate the mathematical formulation of the model decision expression into the Opl language.
minimizeiFfiyi+(iF,jC)cijxij.(37)
The two decision expressions have been placed above and below, to compare and juxtapose the translation so that other models can also be analyzed and translated into the Opl language model.
dexpr float OpeningCost=sum(i in ulabfacilities) fi[i]*y[i];
dexpr float TotalDistance=sum(e in edges) c[e]*x[e];
5) Objective Function
The objective function is that of the UFL problem in that it is to minimize the total sum of opening costs, and the total distance traveled from (ulabcity) to (ulabfacility) location.
Minimize OpeningCost +TotalDistance;
6) Constraints
The constraints are translated from the mathematical formulation of the UFL problem into Opl as a code block subjectto{}. In the constraint code block subjectto{}, the first constraint is placed as a for loop.
subject to {
forall(j in ulabcities: j>=1)
sum (i in ulabfacilities) x[<i, j>]==1;
i in ulabfacilities: i>=1, j in ulabcities: j>=1
sum(i in ulabfacilities) x[<i, j>]==1; x[<i, j>]<=y[i];
}
7) The Data
The CPLEX data file contains the data values which are entered manually. The values of n the number of (ulabfacilities), m the number of ulabcities. It also contains the values of the two arrays, ulabcityLocation and ulabfacilityLocation which host the x andy coordinates of ulabcity locations and ulab collection facility locations. It also contains the values of the array fi in which is entered the opening cost values of the ulab collection facilities.
8) The Input File
The location input is within an array of x and y coordinates ulabcitylocation and ulabfacilityLocation: [<x y> <x y>] can be replaced by GPS Coordinates for example: [< -20.172544 57.4883317 > <-20.165595 57.492625>]. The input coordinates are easier and more convenient to input for the calculation of the UFL problem compared with a distance matrix of the same coordinate points. This makes the algorithm solution user friendly, and usable in a real-world case.
Li shows an example problem where Walmart tasks the k-median problem, Mvere , an American retail giant, to select the best locations to site their retail stores. The aim was to site their 50 stores in Washington state so that the average travelling time for people to their nearest Walmart shop is minimized . This is an example of a real-world case where the average travelling time, distance, is minimized. However, the decision is only concentrated on financial terms, monetized terms . There are other criteria such as competition with other retail stores within the vicinity, the presence of other buildings, residential areas, road density, population density and demographics and other factors that are not put into consideration with location models like Mvere and Jain, Mahdian which may affect the performance of the Walmart store when it is located on any optimally selected facility location by the UFL or k-median problem.
The DEA and TOPSIS method may be used in Location analysis to make an informed decision in such a case as suggested by Li which includes many other criteria. Each of the 50 locations is turned into a DMU and listed as an alternative, its performance as a potential location is evaluated according to the multiple criteria. This involves a lot of data gathering. It also involves a lot of subjective choices and preferences of the decision makers and the community.
Figure 4. Input file. Source Mvere
2.3. Hierarchical Groups DEA Super-efficiency and Group TOPSIS Technique: Application on Mobile Money Agents’ Locations
Multi Criteria Decision Analysis (MCDA) is the umbrella term for methods that deal with multiple criteria, while DEA and TOPSIS are specific tools within that framework. They all contribute to the process of making informed decisions when there are multiple factors to consider . While DEA is a non-parametric method used to evaluate the relative efficiency of DMUs, which are entities that convert inputs into outputs . TOPSIS ranks alternatives based on their distance to both an ideal solution and a non-ideal solution . In location analysis, DEA and TOPSIS were used by Muvingi, Peer , Chen, Li and Zangeneh, Akram.
Hierarchical groups DEA super-efficiency refers to a method used in (DEA) to assess the efficiency of DMUs that have a hierarchical structure, where some DMUs are part of larger groups, for example districts and suburbs .
It extends the concept of super-efficiency, which is used to rank and compare DMUs that are already efficient in the traditional DEA analysis . Super-efficiency methods are used to rank and compare efficient units. DEA determines a "best-practice frontier" by identifying the most efficient DMUs and then compares other DMUs against this frontier . Traditionally, DEA identifies efficient DMUs, meaning those that are operating on the best-practice frontier, and inefficient DMUs, those below the frontier . DEA arises from situations where the goal is to determine the productive efficiency of a system by comparing how well the system converts inputs into outputs, while MCDA models have arisen from the need to analyze a set of alternatives according to conflicting criteria
Instead of the traditional cost-benefit analysis in which all criteria are converted to a monetized term as a common comparison ground like the UFL and k-median problems, this framework incorporates non-monetized considerations into the management planning process, thereby improving the project appraisal process . Both quantitative criteria that can be measured in numerical values objectively and qualitative criteria that can only be gauged subjectively are integrated together to generate comprehensive evaluations for all alternatives. Muvingi, Peer categorize the preferable locations of 16 districts and 49 suburbs as the alternatives. The alternatives are ranked according to criteria such as Distance from CBD (D), Road density (RD), Zoning tariff (ZT), Number of shops (NS), Number of mobile phone users (NMPU), Mobile transfer service users (MTSU) and Population (P) as seen on Figure 5.
Identifying the set of alternatives, A = {a1, a2, an} and the set of criteria, C = {c1, c 2, cq}. Step (1) also provides a direct physical measurement as the consequence of alternative ai on criterion cj for every i = 1,..., n and j = 1,..., q, denoted by m ij, representing the (i, j)-entry of an n × q matrix, called the information (or performance) matrix . The matrix on Figure 7 is transposed compared to the one on Figure 6. The Matrix on Figure 7 shows a Hierarchy of district and suburb alternatives to the left and criterion listed on the top.
Figure 5. Integrated model for the locations of mobile money agents. Source: Muvingi, Peer
Figure 6. DEA performance matrix Source: Chen, Li .
Figure 7. Hierarchical DEA performance matrix. Source: Muvingi, Peer
In Muvingi, Peer , there were three main objectives:
1) Ranking potential suburb and district locations for Mobile Money Agents (MMAs) using seven different SE approaches .
2) Adjust the SE ratings of the suburb locations, termed level 1 DMUs, with the SE ratings of the district locations, termed level 2 DMUs, in which they are grouped .
3) Combine the seven-suburb aggregated SE (SASE) ratings with the TOPSIS method to get a universal ranking of the suburbs and district locations .
Muvingi, Peer sought to incorporate slacks variables directly into the efficiency analysis of DMUs with a hierarchical group structure composed of the suburb and district locations of MMAs. Seven hierarchical DEA SE models that directly incorporated slack variables were developed and applied in the ranking of suburb and district locations . The suburb locations were arranged and grouped in the district locations Noura, Jahanshahloo shows an example of slacks-based measure (SBM) of efficiency of the DMUs,
Noura, Jahanshahloo deal with n DMUs with the input and output matrices
X=(xij)∈Rm×n and Y=(yrj)∈Rs×n, respectively.
They assume that the data set is nonnegative, i.e., X≥0 and Y≥0 the production possibility set ρ is defined as :
(38)
where λ is a non-negative vector in Rn.
A study consider an expression for describing a certain
DMUo= (xo, yo) as:
xo=Xλ+s,(39)
yo=Yλ−s+.(40)
With λ ≥ 0, s≥ 0 and s+ ≥ 0. The vector s∈ Rm and s+ ∈ Rs indicate the input excess and output shortfall of this expression, respectively, and are called slacks.
From the condition X ≥ 0 and λ ≥0, it holds that
xo≥s.
Using s and s+, Noura, Jahanshahloo define an index ρ as follows:
(41)
It can be verified that ρ satisfies the properties (i) units invariant and (ii) monotone decreasing in input/output slacks.
From (84), it holds that:
0 < ρ ≤1.
In an effort to estimate the efficiency of (xo, yo), they formulate the following fractional program SBM in λ, s and s+.
(42)
SBM can be transformed into a linear program using the Charnes–Cooper transformation in a comparable way to the CCR model.
The analysis of the level 1 and 2 SE ratings is derived from the seven models implemented for DMUs previously assumed to be homogeneous and not arranged in a group structure . To ensure that those models are applicable to the SE evaluation of DMUs in a two-level hierarchical group structure, additional constraints were added:
Figure 8. DSE-1 ratings. Source: Muvingi, Peer DSE-1 ratings. Source: Muvingi, Peer
DSE-1 ratings .
A study introduced an improved Super SBM model that can handle negative data, which is an improvement from the traditional super SBM models. They propose a non-radial SE model capable of ranking efficient DMUs .
Figure 9. DSE-2 ratings. Source: Muvingi, Peer DSE-2 ratings. Source: Muvingi, Peer
Muvingi, Peer consider a modified slacks based SE measure in the presence of negative data. The SE methodology, for example that of Bajec, Tuljak . is crafted to manage data uncertainties, while the multi-level nature of mobile money agent networks in Muvingi, Peer can be somewhat complex.
Figure 10. DSE-5 ratings. Source: Muvingi, Peer . DSE-5 ratings. Source: Muvingi, Peer .
The Muvingi, Peer study implemented the TOPSIS method to obtain a clear ranking of all the suburb locations through the incorporation of all the seven DSE ratings . Their ranking results revealed that most of the highly-rated suburb areas are located in low density areas, while the opposite was observed in the TOPSIS ranking of district locations . They conjectured that a complete ranking of the DMUs through TOPSIS was achieved for all district and suburb locations . The adoption of the TOPSIS method provided rankings derived from the ratings of seven diversified SE approaches, which ensured that the merits and demerits of each DEA SE model were considered in the final ranking of DMUs . The TOPSIS method refined the ranking by comparing each agent’s performance against an ideal solution, thereby overcoming the limitations of traditional DEA in distinguishing between efficient units.
Figure 11. DSE ratings and rankings. Source: Muvingi, Peer .
Figure 12. TOPSIS rankings. Source: Muvingi, Peer.
Muvingi, Peer used the Principal Component Analysis (PCA) method and the Spearman correlation analysis to verify and validate the TOPSIS results. They found the accuracy of the district location TOPSIS ranking approach to be moderate, at the rate of 61 percent on the observed suburb locations. Mvere used the ESMVERE-R-hamming k-median clustering method to infer opinions of whether there would be a need to introduce a new collection system for ULABs to which 55 percent of the respondents confirmed the need for a new ULAB collection system . However, the study, Muvingi, Peer , contributed to the literature focused on comparing TOPSIS and PCA, as most of the work is on the comparison of DEA and PCA. While Mvere introduced an original, completely new and innovative way of analyzing questionnaire data.
The main contribution of the study Muvingi, Peer includes:
1) It provides insight to bank and Mobile Money Operators (MMOs) regulators into how the characteristics of locations are related to the use of mobile money, represented by MTSU.
2) Regulators can use the efficiency analysis of the locations in the study as a proxy measure for potential financial inclusion in a given location.
3) The study provided a bench-marking analysis of potential MMA locations, allowing banks to strategically locate their branches depending on the influence of MMAs locations on their operations.
4) The location performance results guide MMOs for MMA network scaling.
3. Conclusion
In situations which demand urgent coverage of the entire population indiscriminately, for example situations where there is a national disaster or a disease outbreak, to cover the entire population, the UFL problem or the k-median problem would be a better choice technique. One of the most complex decisions in such a real-world case would be to determine what constitutes a facility, how much a facility would cost to build and what would represent a client . The information would determine how many potential facility locations sites and client locations sites would be in an instance, thus the size of an instance.
In such situations, the ESMVERE CPLEX algorithms would prove to be better solutions because of them being user friendly on the input and speedy in how they process the data. Lead poisoning is a danger, although it is a subtle and long-term danger it is still of paramount importance that every community has safe disposal facilities near them. Thus, the disposal of such hazardous waste would not involve preference or competition in a normal scenario.
This acutely differs from situations where the goal is to attract certain types of customers. This situation requires more information about the potential site. Mobile money services and banking services in general involve preferences. The preference of the decision makers in selecting places that maximize their value and minimize their costs, places that are away from competition, and places that are convenient and easily accessible to their customers. On the other hand, the customer usually has a choice in selecting a service from a particular location. They may not always seek a facility which is near them, but they may choose to seek their services in places that are less congested, easier to reach, where the mobile phone network responds faster etc. These are conditions where the DEA and TOPSIS technique would be an ideal solution.
In the case of Walmart , for 50 stores sitting in Washington state, the ESMVERE algorithms k-median version, can be more useful in the decision because of speediness to a solution which covers the whole state. However, in such a case, to a greater extent, the Hierarchical groups DEA and group TOPSIS technique will afford more information which could result in better decisions. This is a situation in which the two techniques could be complimentary and used together.
The ESMVERE CPLEX k-median problem solver algorithm is used to determine the optimum or near optimum locations which minimize the average travel time of every customer in the state of Washington to a Walmart shop of maybe about 50 locations. Then the Hierarchical groups DEA and group TOPSIS technique will determine and rank the best out of those 50 locations based on relevant criteria. This could mean ranking the performance of different values of k such as k= 40 or k=50 or k=60 to determine the best value for k which covers the entire population . The two techniques used in conjunction could mean a more rigorous and time demanding data gathering process but better decision making.
The dual fitting and factor revealing LP technique remains complex to implement in such a real-world case as that of Mvere because of the input of a distance matrix. To measure 22464 location distances accurately would be tedious, this solution compared to the input of 156 GPS Coordinate facility locations and 144 GPS coordinate client locations which proves the usefulness of the new solution invented by Mvere . Although the algorithm Jain and Mahdian on the Tables 1, 2, 3, 4 and Figures 1 and 2, is well formulated in theory, more needs to be done to achieve its use to solve real world problems. Future work in the area might mean reformulation of the algorithm to adopt the methodology of which include the Euclidean Distance formula and GPS coordinate input,
While Muvingi, Peer offers a hybrid evaluation framework that integrates hierarchical groups DEA with group TOPSIS to assess the efficiency of mobile money agent’s locations in Zimbabwe. The detailed methodology and empirical application within the Zimbabwean context offer valuable insights for evaluating service delivery in digital financial systems.
The ESMVERE CPLEX UFL problem solver algorithms can extends to a broader range of facility location challenges across different industries and sectors. The algorithm's robust scalability and solution quality make it suitable for various UFL location problems where decision-makers must identify optimal facility sites to minimize costs or maximize service efficiency without capacity constraints.
Beyond hazardous waste management, the ESMVERE UFL algorithm can address applications such as:
1) Health Care Facility Location Optimizing the placement of clinics, hospitals, or emergency response centers to improve accessibility and reduce response times in urban or rural areas.
2) Retail and Distribution Centers Determining the best locations for warehouses or retail outlets to balance logistics costs and market coverage efficiently.
3) Public Service Infrastructure Siting facilities like schools, fire stations, or public offices where service coverage and proximity to populations are critical.
4) Telecommunications Networks Facilitating the positioning of network hubs or base stations to ensure optimal coverage with minimal deployment costs.
These potential applications leverage the algorithms capability to handle large problem instances, with complex cost structures and constraints, enabling practical deployment in varied operational contexts. The balanced integration of sophisticated optimization techniques within the ESMVERE framework enables the algorithms to maintain scalability without compromising on solution accuracy.
In conclusion, this systematic review successfully highlights the differences and similarities among three significant location science methods. Each method offers different real world applications, as demonstrated through practical case studies in mobile banking and hazardous waste management. To enhance the practical utility of these methods, further exploration of their implications is essential. Future research should focus on integrating these techniques with emerging technologies to improve decision making processes in various sectors
Abbreviations

CBD

Central Business District

DEA

Data Envelopment Analysis

DMU

Decision-Making Units

GPS

Global Positioning System

HYB

Hybrid Heuristics

ILS

Iterated Local Search

MCDA

Multi Criteria Decision Analysis

MDAS

Multi-Start Descending Algorithm with Screening

MMA

Mobile Money Agents

MMO

Mobile Money Operators

Np-hard

Non-Deterministic Polynomial Time

Hard

Opl

Optimization Programming Language

P

Polynomial Time

PCA

Principal Component Analysis

TOPSIS

Technique for Order Performance (preference) by Similarity to Ideal Solution

SASE

Suburb Aggregated Super Efficiency

SE

Super Efficiency

SBM

Slacks Based Measure

UFL

Uncapacitated Facility Location

ULAB

Used Lead Acid Battery

Acknowledgments
I would like to thank my parents, My Mum who has made all my research efforts possible., My whole family, my darling, my colleagues. The people of Zimbabwe, the people of Mauritius, the people of Africa, and all the people of the world. I would like to thank My lord and Savior JESUS CHRIST for turning my dream into a reality.
Author Contributions
Emmanuel Siyawo Mvere is the sole author. The author read and approved the final manuscript.
Conflicts of Interest
The authors declare no conflicts of interest.
Appendix
The ESMVERE CPLEX UFL Problem Solver Model
The ESMVERE CPLEX UFL problem solver algorithm is a software invention that solves the UFL problem. It is a tool that aids, and guides management in making sound decisions on where to locate facilities. It aids management in the decision of where the best places are, lowest cost places, the minimum distance places to locate facilities considering that there are opening costs and the Euclidean distances, direct line distances of GPS coordinate points?
The CPLEX Opl Code
int n=...; //number of facilities
int m=; //number of clients declared in data files range ulabfacilities=1..n;                 
range ulabcities=1..m; int fi[ulabfacilities]=...;
tuple Location {float x;float y;}
tuple edge {int i;int j;}
setof (edge) edges={<i,>| i in ulabfacilities, j in ulabcities };//defines an edge
float c[edges]; //An Array of each edge on the plane measured separately
Location ulabcityLocation [ulabcities];
Location ulabfacilityLocation [ulabfacilities];
execute {
function getDistance (ulabcity1, ulabfacilitycity1) { //Euclidean distance between each facility and client
return opl.sqrt(opl.pow(ulabcity1.-ulabfacility1.x,2)+opl.pow(ulabcity1.y
-ulabfacility1.y,2)) }
for(var i inulabfacilities) {
ulabfacilityLocation[i].x;
ulabfacilityLocation[i].y;
}
for(var j inulabcities) {
ulabcityLocation[j].x;
ulabcityLocation[j].y;
}
for(var e in edges){ //distance between each edge “heuristic”
c[e]=getDistance(facilityLocation[e.i],cityLocation[e.j])
} }
dexpr float OpeningCost=sum(i in ulabfacilities) fi[i]*y[i];
dexpr float TotalDistance=sum(e in edges) c[e]*x[e];
Minimize OpeningCost +TotalDistance; //objective function
subject to { //constraints from p-median
forall(j in ulabcities: j>=1)
sum (i in ulabfacilities) x[<i, j>]==1;
forall(i in ulabfacilities, j in ulabcities)
x[<i,j>]<=y[i];
sum(i in ulabfacilities)
}
Figure 13. Input File. Input file. Source Mvere
Input File
In this case the input are x and y coordinates [<x y> <x y>] can be replaced by
GPS Coordinates [< -20.172544 57.4883317 > <-20.165595 57.492625>] compare with a distance matrix of the same coordinate points.
Figure 14. k-median Optimum Solution. Source Mvere , Maudho .
Figure 14 shows the final near optimum solution which hosts nineteen facilities.
References
[1] E. S. Mvere. (2025) The Near Optimal Siting of Hazardous Waste (Used Lead Acid Battery (ULAB)) Collection Facilities in The Republic of Mauritius Using the Esmvere Cplex k-median solver algorithm. In K. Madhu (Eds,), Advances in Numerical Analysis and Applications. Vol 1, pp, 16-46, ISBN: 978-81-958975-6-8.
[2] E. S. Mvere.(2025) The Near Optimal Siting of Hazardous Waste (Used Lead Acid Battery (ULAB)) Collection Facilities in The Republic of Mauritius Using the ESMVERE Cplex UFL problem solver algorithm. In K. Madhu (Eds,), Advances in Numerical Analysis and Applications. Vol 1, pp, 47-80, ISBN: 978-81-958975-6-8.
[3] Muvingi, J., Peer, A. A. I., Hosseinzadeh Lotfi, F., & Ozsoy, V. S. (2023). Hierarchical groups DEA super-efficiency and group TOPSIS technique: Application on mobile money agents’ locations, Expert Systems with Applications, Volume 234, 0957-4174,
[4] Jain, K., Mahdian, M., Markakis, E., Saberi, A. & Vazirani, V. V. (2003). Greedy facility location algorithms analyzed using dual fitting with factor-revealing Lp. J. ACM.
[5] Chen, Y., Li, K. W., Xu, H.(2009) A DEA-TOPSIS method for multiple criteria decision analysis in emergency management. J. Syst. Sci. Syst. Eng. 18, 489–507.
[6] Mvere, E. S (2025). A Systematic Review Of The Esmvere Algorithms, Which Are Contributions In The Field Of Uncapacitated Facility Location Problems And K‐Median Problems International Journal of Science Academic Research, Vol. 06 No. 2, 2025, pp. 9352-9362
[7] Jain, K. & Vazirani, V. (1999). Primal-dual approximation algorithms for metric facility location and k-median problems. in ‘Proceedings of the 40th Annual Symposium on Foundations of Computer Science’. FOCS ’99. p. 2.
[8] Faiz, S. & Krichen, S. (2012) Geographical Information Systems and Spatial Optimization Vol 1, pp, 176, ISBN 9780429072079 CRC Press
[9] Harder, H. D. (2023) Exact Algorithm or Heuristic?
[10] Li, S. (2014). Approximation Algorithms for Network Routing and Facility Location Problems. PhD thesis. Princeton University.
[11] Svensson, O. (2009). Approximability of Some Classical Graph and Scheduling Problems. PhD thesis. University of Lugano.
[12] Swamy, C. (2004). Approximation Algorithms for Clustering Problems. PhD thesis. Cornell University.
[13] Woeginger, G. J. (2003). Exact Algorithms for NP-Hard Problems: A Survey. In: Jünger, M., Reinelt, G., Rinaldi, G. (eds) Combinatorial Optimization — Eureka, You Shrink!. Lecture Notes in Computer Science, vol 2570. Springer, Berlin, Heidelberg.
[14] IBM (2025) Decision Optimization at a glance
[15] Charles, V., Aparicio, J., Zhu, J.(2019) The curse of dimensionality of decision-making units: A simple approach to increase the discriminatory power of data envelopment analysis, European Journal of Operational Research, Volume 279, Issue 3, Pages 929-940, ISSN 0377-2217,
[16] Ramdania, D. R., Manaf, K., Junaedi, F. R., Fathonih, A., Hadiana, A,(2020) TOPSIS Method on Selection of New Employees’ Acceptance,"6th International Conference on Wireless and Telematics (ICWT), Yogyakarta, pp. 1-4,
[17] Li, S.(2013) A 1.488 approximation algorithm for the Uncapacitated facility location problem, Information and Computation, Vo 222, 2013, Pages 45-58,
[18] Nguyen, K. T.,(2019) Primal-Dual Approaches in Online Algorithms, Algorithmic Game Theory and Online Learning. Operations Research [math. OC]. Université Paris Sorbonne. fftel-02192531f.
[19] Ebrahimnejad, A.,(2012) A primal-dual method for solving linear programming problems with fuzzy cost coefficients based on linear ranking functions and its applications," International Journal of Industrial and Systems Engineering, Inderscience Enterprises Ltd, vol. 12(2), pages 119-140.
[20] Chu, C., Antonio, J., (1999) Approximation Algorithms to Solve Real-Life Multicriteria Cutting Stock Problems Operations Research 47: 4, 495-508.
[21] Wikipedia contributors. (2025, April 25). Approximation algorithm. In Wikipedia, The Free Encyclopedia. Retrieved 23: 05, May 25, 2025, from
[22] Maria, G. L., Hua, W., Henry, W.,(2009). A Stable Primal-Dual Approach for Linear Programming under Nondegeneracy Assumptions. Computational Optimization and Applications. 44. 213-247.
[23] Mvere, E. S (2025) An Analysis of the Used Lead Acid Battery (ULAB) Collection Network Mauritius Using the ESMVERE- R-Hamming-k-median Clustering Method., International Journal of Science Academic Research, Vol. 06 No. 1, 2025.
[24] Balinski, M. (1966). On finding integer solutions to linear programs In: Proceedings of the IBM Scientific Computing Symposium on Combinatorial Problems, pp. 225-248.
[25] Hakimi, S. L., (1964). Optimum locations of switching centers and the absolute centers and medians of a graph. Operations Research, 12(3), pp. 450-459.
[26] Cacioppi, P.& Watson, M. (2014). A Deep Dive into Strategic Network Design Programming: OPL CPLEX Edition. Northwestern University Press.
[27] Mistry, S (2019) Optimization-Facility-Location [source code].
[28] Srinivasan, G. (2012). Mod-08 Lec-32 Location problems -- p median problem, Fixed charge problem.
[29] Cornuejols, G., Nemhauser, G & Wolsey, L (1984). The Uncapacitated Facility Location Problem. Cornell University
[30] Caceres, H. (2014). CPLEX Seminar Getting started with CPLEX Studio (part 2).
[31] Zangeneh, M., Akram, A., Nielsen, P., Keyhani, A., & Banaeian, N. (2015). A solution approach for agricultural service center location problems using TOPSIS, DEA and SAW techniques. In P. Golinska, & A. Kawa (Eds.), Technology management for sustainable production and logistics (pp. 25–56). Springer, Berlin, Heidelberg,
[32] Muvingi, J., Peer, A. A. I., Hosseinzadeh Lotfi, F., & Ozsoy, V. S. (2022). Hierarchical groups DEA cross-efficiency and TOPSIS technique: An application on mobile money agents locations. International Journal of Mathematics in Operational Research, 21(2), 171–199.
[33] Labijak-Kowalska, A., Kadziński, M., & Mrozek, W. (2023). Robust Additive Value-Based Efficiency Analysis with a Hierarchical Structure of Inputs and Outputs. Applied Sciences, 13(11), 6406.
[34] Noura, A. A.,. Jahanshahloo, H. F., G. R., Rashidi, S. F. (2011) Super-efficiency in DEA by effectiveness of each unit in society, Applied Mathematics Letters, Volume 24, Issue 5, Pages 623-626, ISSN 0893-9659.
[35] Bajec, P., Tuljak-Suban, D., & Zalokar, E. (2021). A Distance-Based AHP-DEA Super-Efficiency Approach for Selecting an Electric Bike Sharing System Provider: One Step Closer to Sustainability and a Win–Win Effect for All Target Groups. Sustainability, 13(2), 549.
[36] Li, S.(2013) Approximation Algorithms for Facility Location Problems and Network Routing Problems.
[37] Mahdian, M., and Spielman, D. A. (2004). Facility location and the analysis of algorithms through factor-revealing programs. Ph. D. Dissertation. Massachusetts Institute of Technology, USA. Order Number: AAI0806252.
[38] Jain, K., Mahdian, M. and Saberi, A. (2002). A new greedy approach for facility location problems. in ‘Proceedings of the thirty-fourth
[39] Guha, S. and Khuller, S. (1998). Greedy strikes back: Improved facility location algorithms. Journal of Algorithms p. pp. 649–657.
[40] Ahmadian, S. (2017). Approximation Algorithms for Clustering and Facility Location Problems. PhD thesis. University of Waterloo.
[41] Peidro, D., Martin, X., Panadero, & Juan, A. (2024). Solving the Uncapacitated facility location problem under uncertainty: a hybrid tabu search with path-relinking simheuristic approach
[42] Chen, H., Murray, A. T. & Jiang, R. Open-source approaches for location cover models: capabilities and efficiency. J Geogr Syst 23, 361–380 (2021).
[43] Dam, T., Ta, A. T., and Mai, T.(2003) Robust maximum capture facility location under random utility maximization models, European Journal of Operational Research, Vol 310, Issue 3, pp 1128-1150.
[44] Tong D., Murray, A. T.(2009) Maximizing coverage of spatial demand for service. Pap Reg Sci 88(1): 85– 97.
[45] Maudho, A. (2023). ArcGIS Cartography Open Data Portal, Ministry Of Information Technology Communication and Innovation Mauritius.
[46] Hamdani, and Wardoyo, R. (2016). The complexity calculation for group decision making using TOPSIS algorithm. In AIP conference proceedings (Vol. 1755, No. 1, p. 070007). AIP Publishing LLC.
[47] Muvingi, J., Peer, A. A. I., Jablonský, J., Azizi. H., and Lotfi, H. F., (2024). Hierarchical fuzzy DEA model with double frontiers combined with TOPSIS technique: application on mobile money agents locations, OPSEARCH, Springer; Operational Research Society of India, vol. 61(3), pages 1154-1191, September.
Cite This Article
  • APA Style

    Mvere, E. S. (2025). A Systematic Review, Comparison and Juxtaposition of the (Dual Fitting and Factor Revealing LP Technique), (the ESMVERE CPLEX Algorithms) and the (Data Envelopment Analysis and TOPSIS Technique). American Journal of Operations Management and Information Systems, 10(1), 13-37. https://doi.org/10.11648/j.ajomis.20251001.12

    Copy | Download

    ACS Style

    Mvere, E. S. A Systematic Review, Comparison and Juxtaposition of the (Dual Fitting and Factor Revealing LP Technique), (the ESMVERE CPLEX Algorithms) and the (Data Envelopment Analysis and TOPSIS Technique). Am. J. Oper. Manag. Inf. Syst. 2025, 10(1), 13-37. doi: 10.11648/j.ajomis.20251001.12

    Copy | Download

    AMA Style

    Mvere ES. A Systematic Review, Comparison and Juxtaposition of the (Dual Fitting and Factor Revealing LP Technique), (the ESMVERE CPLEX Algorithms) and the (Data Envelopment Analysis and TOPSIS Technique). Am J Oper Manag Inf Syst. 2025;10(1):13-37. doi: 10.11648/j.ajomis.20251001.12

    Copy | Download

  • @article{10.11648/j.ajomis.20251001.12,
      author = {Emmanuel Siyawo Mvere},
      title = {A Systematic Review, Comparison and Juxtaposition of the (Dual Fitting and Factor Revealing LP Technique), (the ESMVERE CPLEX Algorithms) and the (Data Envelopment Analysis and TOPSIS Technique)
    },
      journal = {American Journal of Operations Management and Information Systems},
      volume = {10},
      number = {1},
      pages = {13-37},
      doi = {10.11648/j.ajomis.20251001.12},
      url = {https://doi.org/10.11648/j.ajomis.20251001.12},
      eprint = {https://article.sciencepublishinggroup.com/pdf/10.11648.j.ajomis.20251001.12},
      abstract = {The objective of the present paper is to compare three location science methods. Two of the solutions discussed are practical real-world case studies, while the other method is used to reveal the approximation factor of an algorithm. We conduct an in-depth analysis of the originality, similarities, and differences of the three methods to unravel clarity on the conditions where each method may be applied, or where one method might be preferred over another. (1) The Data Envelopment Analysis (DEA)-Technique for Order Performance (preference) by Similarity to Ideal Solution (TOPSIS) technique is usable in many other areas of management science besides location analysis, however it is used in this case to guide the decision of selecting the best locations among alternatives. Hierarchical groups DEA super-efficiency and group TOPSIS technique is applied on mobile money agents’ locations in Harare Zimbabwe. (2) The ESMVERE algorithms introduce novel reformulations of the Uncapacitated Facility Location (UFL) and k-median problems which replace traditional distance matrices with Global Position System (GPS) based Euclidean Distance calculations. They are platform dependent solutions based on the CPLEX Optimization programming language (Opl). The solutions are used to solve a real-world problem which involve the near optimal siting of Hazardous Waste, Used Lead Acid Battery (ULAB), collection facilities in the Republic of Mauritius. (3) The Dual fitting with factor-revealing Linear Program (LP) technique is a location science approximation algorithm solution, used to analyze greedy algorithms that select facilities and connect the clients based on the dual solution information. The thrust of this review is to analyze the complexity, compare, contrast, and discuss the three location science techniques so as to classify them based on their purpose and area of application. Hence this review analyzes three papers and summarizes the main contributions of the three papers.
    },
     year = {2025}
    }
    

    Copy | Download

  • TY  - JOUR
    T1  - A Systematic Review, Comparison and Juxtaposition of the (Dual Fitting and Factor Revealing LP Technique), (the ESMVERE CPLEX Algorithms) and the (Data Envelopment Analysis and TOPSIS Technique)
    
    AU  - Emmanuel Siyawo Mvere
    Y1  - 2025/08/01
    PY  - 2025
    N1  - https://doi.org/10.11648/j.ajomis.20251001.12
    DO  - 10.11648/j.ajomis.20251001.12
    T2  - American Journal of Operations Management and Information Systems
    JF  - American Journal of Operations Management and Information Systems
    JO  - American Journal of Operations Management and Information Systems
    SP  - 13
    EP  - 37
    PB  - Science Publishing Group
    SN  - 2578-8310
    UR  - https://doi.org/10.11648/j.ajomis.20251001.12
    AB  - The objective of the present paper is to compare three location science methods. Two of the solutions discussed are practical real-world case studies, while the other method is used to reveal the approximation factor of an algorithm. We conduct an in-depth analysis of the originality, similarities, and differences of the three methods to unravel clarity on the conditions where each method may be applied, or where one method might be preferred over another. (1) The Data Envelopment Analysis (DEA)-Technique for Order Performance (preference) by Similarity to Ideal Solution (TOPSIS) technique is usable in many other areas of management science besides location analysis, however it is used in this case to guide the decision of selecting the best locations among alternatives. Hierarchical groups DEA super-efficiency and group TOPSIS technique is applied on mobile money agents’ locations in Harare Zimbabwe. (2) The ESMVERE algorithms introduce novel reformulations of the Uncapacitated Facility Location (UFL) and k-median problems which replace traditional distance matrices with Global Position System (GPS) based Euclidean Distance calculations. They are platform dependent solutions based on the CPLEX Optimization programming language (Opl). The solutions are used to solve a real-world problem which involve the near optimal siting of Hazardous Waste, Used Lead Acid Battery (ULAB), collection facilities in the Republic of Mauritius. (3) The Dual fitting with factor-revealing Linear Program (LP) technique is a location science approximation algorithm solution, used to analyze greedy algorithms that select facilities and connect the clients based on the dual solution information. The thrust of this review is to analyze the complexity, compare, contrast, and discuss the three location science techniques so as to classify them based on their purpose and area of application. Hence this review analyzes three papers and summarizes the main contributions of the three papers.
    
    VL  - 10
    IS  - 1
    ER  - 

    Copy | Download

Author Information
  • Mechanical and Production Engineering Department, University of Mauritius, Moka, Mauritius

    Biography: Emmanuel Siyawo Mvere earned his MPhil in Operational Research, in the Department of Mechanical and Production Engineering, Faculty of Engineering of the University of Mauritius. He earned his MSc Applied Statistics with Operational Research in the Department of Applied Mathematical Sciences, School of Innovative Technologies and Engineering, of the University of Technology Mauritius. He earned his BTech (honours) Degree in Electronic Commerce in the School of Business and Management Sciences of the Harare Institute of Technology, Zimbabwe. He earned his Certificate in Applied Information Technology specializing in Network Engineering in the School of Technology, University of Zimbabwe, Inventor of software solutions: The ESMVERE CPLEX UFL problem solver algorithm, The ESMVERE CPLEX k-median problem solver algorithm, The ESMVERE CPLEX k-UFL problem solver algorithm and the ESMVERE-R-Hamming-k-median clustering data analysis method, among many others. Recently published by invite, 2 book chapters in an editorial book: Advances in Numerical Analysis and Applications.

  • Abstract
  • Keywords
  • Document Sections

    1. 1. Introduction
    2. 2. Main Contents
    3. 3. Conclusion
    Show Full Outline
  • Abbreviations
  • Acknowledgments
  • Author Contributions
  • Conflicts of Interest
  • Appendix
  • References
  • Cite This Article
  • Author Information