< d f f f d e cce c c e d c ddcfede c c e dad c ed e > no. vol. al-rafidain engineering accepted th march submitted th may a computer expert system for analysis and control of water hammer problems dr. rasul m. khalaf prof. dr. alaa h. kadoury college of engineering college of engineering university of tikrit univerity of al-mustansiria wisam j. al-hilo college of engineering university of al-mustansiria abstract the current system called expert system for analysis and control of water hammer problems “esacwhp” is developed to help engineers in the design and analysis of water hammer problems using the characteristic method with aid of the programming language called visual basic. results obtained from the developed of “esacwhp” show good agreement with that solved by traditional lengthy methods, in addition it is capable of handling more variables. the system is recommended to be used for analysis and design. .. . )characteristic method ( )windows ()visual basic ( . . no. vol. al-rafidain engineering . introduction expert systems are computer programs that contain knowledge in a specific domain. the expert system often uses this knowledge to perform tasks a human can do, but with a longer time. expert system has become commercially important because it provides the way of archiving expertise and making it routinely available where it is needed. since early eighties, researchers have developed an entirely new kind of tools, a computer that serves as an assistant whose skills include finding reasonable solutions to problems for which there may be no hard and fast “right” answers. the “expert” computer system uses extensive experience-based knowledge of a subject to guess intelligently in the same way as a human expert does. before computer analysis, the general equations describing water hammer in pipeline system are simplified in some manner to permit solution by arithmetic, graphical… etc. matching the boundary condition at pumps and turbines is at best, difficult and understood by relatively few engineers. modern analysis techniques, including numerical methods of solving partial differential equations, are brought within the capability of solving a wide range of water hammer problems. to protect the piping system against the extreme pressure developed during the transient state, it can be designed with a liberal factor of safety to withstand these pressure changes. such a design will be uneconomical and various devices are employed to control these changes. two types of devices are available. the first, such as surge tank, air vessels and bypasses, reduces the rate of net change in initial conditions. the second, such as safety valve, pressure regulating valves and air inlet valves, restricts the minimum or maximum pressure developed during the transient state. a brief description of the two types of devices may be found in ref. [ ]. the designed expert system, which is constructed to control, and analyze water hammer, is called “expert system for analysis and control of water hammer problems” and in abbreviation,can be written as “esacwhp”. . theoretical aspects befor introducing a brief discription of general equation for water hammer problem, a basic architecture of an expert system is viewed. . the structure of an expert system a human expert uses knowledge and reasoning to arrive at conclusions. similarly, an expert system relies on knowledge and performance reasoning. the reasoning carried out in expert system attempts to mimic human experts in combining pieces of knowledge. thus, the structure or architecture of an expert system partially resembles how a human expert performs. thus, there is an analogy between an expert and an expert system. no. vol. al-rafidain engineering the basic architecture of an expert system consists of three parts: the knowledge base, the inference engine (inference mechanism) and the working memory,as shown in fig.( ) ,[ ]. a. the knowledge base (kb) the knowledge base is the medium through which a human expert’s knowledge is made available to the computer. it contains general problem solving knowledge as well as expert knowledge about how to solve problems or how to interact with the user and it is mostly built into the way the inference engine operates. b. the working memory it contains information that the system has received about the problem at hand. in addition, any information the expert system derives about the problem is stored in the working memory. c. the inference engine (inference mechanism) the inference engine combines facts and rules to arrive at conclusions. but if a large number of facts and rules are matched then how can the systems choose the right facts or rules and the wrong choice may set the system on wild goose chase where most of the reasoning is of little assistance in performing the task. thus, the system needs method of inference that can select which rule can be applied at each step in the reasoning process. in performing inference, the inference engine tries to establish the truth or facility of a statement called a goal. a goal is a fact whose truth-value is to be determined [ ]. . water hammer considerations the characteristic method is utilized a special property of hyperbolic partial differential equations to find their numerical solutions. for a system of hyperbolic partial differential equations, there are two characteristic directions in the s-t plane (see fig. ) in which the integration of the partial differential equations is reduced to the integration of a system of ordinary differential equations. the advantages of this method are: it is a method of solution which allows the direct inclusion of friction losses, it offers ease in handling the boundary conditions and in the programming of complex systems, it is a general method, i.e. the program once written and it can be used for analyzing different piping systems having the same boundary conditions, and the transient state conditions, obtained by using this method are close to the actual situation. the restrictions in this method are the flow must be one-dimensional, the wave speed is constant during the transient state, and the time increment is chosen accordingly to satisfy the stability conditions [ ]. the general equations for water hammer by characteristics method are: no. vol. al-rafidain engineering rrllrlrlrlp vvvv d tf vvt a g hh a g vvv sin ……….( ) rrllrlrlrlp vvvv d tf g a vvtvv g a hhh sin ……....( ) the boundary conditions and control devices can be found in refs.[ ] and [ ] . design and development of esacwhp the expert system is constructed as a skeleton of expert system and as a base of information about water hammer using a visual basic technique.this system is designed to be simple and easy to handle, using two system units (si & english unit). in order to provide basic information for the user, tables of elasticity, density…etc, for liquids at different temperatures are included in the system. the expert system contains aid instruments and auxillary softwares (wave speed calculations, graphical representation …etc) to expand the information about water hammer. . esacwhp properties the properties of the current expert system are described by the following steps: - esacwhp depends on the information bases that essentially related to the water hammer theory .its programs are designed initially using fortran language and changed later to the basic language in order to be properly interactive with visual basic program. - esacwhp displays the questions through out ( ) windows or more. several windows are designed and categorized to fit the most general and partical problems. esacwhp is written with more than program steps , as shown in figs. ( ) and ( ). - the question windows related to the required information are designed to investigate the aims in article ( ) above. - conclusion tool of esacwhp tells the users about the errors that may be committed by mistake in order to overcome them. it gives warnings to avoid the mistakes, as well as it corrects them whensoever happened. - the conclusion tool leads the users to achieve the aims followed in articles ( , ) using the bases of information followed in article ( ). - when all the calculations and operations are achieved throughout the system without mistakes, the required information would display on windows according to their specific locations in the expert system. no. vol. al-rafidain engineering . esacwhp operation when [vertical pump] was choosen and clicking [ok], a related window will be displayed ,as shown in fig. ( ). when click [surge tank], its window will at the start of the esacwhp, a window is displayed ,as shown in fig. ( ), containing the name of the program, the producer, and its version. there is a clause “press any key” at the bottom of the window indicating that the program is ready when any key would be pressed. after that, the main window of the system displays ,as shown in fig. ( ), this window contains the menu bar and the boundary conditions of the system, as listed below: menu bar: - file - wave speed - graph - setting - help boundary conditions: - pump station - surge tank - air chamber - surge valve - dead end - network - valve - atmospheric discharge when click [wave speed] a corresponding window displays, as in fig. ( ), this window is used to calculate wave speed in fluids. when click [graph] a related window displays, as in fig. ( ),this window presents a separated software called wham linked to esacwhp to explain the instant variation of h.g.l. and the discharge for a pipe with a valve installed at the downstream and a reservoir at the upstream, this window includes five options for a valve to be selected, as shown in fig. ( ). when click [pump station], its window will be displayed ,as shown in fig.( ), containing two types of pump: horizontal pump vertical pump when [horizontal pump] was choosen and associating with click [ok], a specific window will be displayed ,as shown in fig. ( ). be displayed, as shown in fig.( ). when click[air chamber], its window will be displayed,as shown in fig. ( ). when click [surge valve], its window will be displayed, as shown in fig. ( ). when click [dead end] , its window will be displayed in order to analyze the flow from a reservoir in which the downstream pipe contains dead end or closed valve, as shown in fig. ( ). no. vol. al-rafidain engineering when click [networks], a corresponding window will be displayed in order to analyze the pipe networks (pipes branch) in each position whereas the user selects, as shown in fig. ( ). when click [valve], its window will be displayed to analyze the flow from a reservoir in which the downstream pipe contains a valve, as shown in fig.( ). when click [atoms. discharge], its window will be displayed, as shown in fig.( ). there are many applications for esacwhp in order to analyze the water hammer problems in hydraulic systems.the results obtained from these applications that can be found in ref. [ ], were verified by traditional lengthy method and the tests show good agreement. . capabilities of esacwhp esacwhp used in this study, is discribed by the followings: - esacwhp reduces the time and helps the hydraulic engineers to analyze the network. - esacwhp is easy and clear to use by engineers. besides it warns the user about the faults throughout the input or within the analysis. - esacwhp can save the results in separate file in order to be used later for further analysis. - esacwhp accepts the future development and is capable of further modifications and improvement to solve other problems which are not considered by esacwhp. - esacwhp contains a branched programs which work as clarification tool for water hammer recycle. . conclusions and recommendations from this research context and from the development of “ esacwhp ”, the followings could be concluded: - a number of published and solved problems are analyzed using the expert system “esacwhp”, the obtained results show a good agreement.for more details see ref.[ ] . - the expert system gives solutions for most of the main problems caused by water hammer phenomenon, such as pumps power failure and sudden closing of valve. - esachwp is a friendly software and can be handled by an engineer with a medium level of computer knowledge. the following modifications can be recommended for the expert system: - the expert system can be modified to include the analysis of a system at the beginning of the pump operation, because of the expected occurrence of changes in the pressure, (before reaching steady state conditions). no. vol. al-rafidain engineering - the expert system can be modified to include the analysis of a system in case of partial operation of pumping sets. - the software “ excel ” can be used to support the expert system “esacwhp”. this makes it able to represent the results in a graphical fashion. references [ ]. al-hilo ,w. j.“ a computer system for analysis and control of water hammer problems ” m.sc. thesis , depart. of environmental engineering, al- mustansiria university , . [ ]. beerel, and annabel, c., “expert system strategic implications and applications ”, ellis horwood ltd., u.k., . [ ]. chaudhry, m.h., “applied hydraulic transients ”, st edition, van nostrand reinhold company, new york, . [ ]. hart, a., “knowledge acquisition for expert systems ” kogan page ltd., nd edition, london, . [ ]. tullis, j.p., “hydraulics of pipelines, pumps, valve, cavitation, transients”, st edition, john wiley & sons, new york, . [ ]. watters, g.z., “analysis and control of unsteady flow in pipeline”, nd edition, butter worth, chicago, california, . no. vol. al-rafidain engineering t s s s s s n s n+ t t p p p p n p n+ c+ c - figure ( ) the characteristic grid for a typical pipe. figure ( ) expert system architecture. expert system program inference engine working memory user output data input data knowledge engineer no. vol. al-rafidain engineering figure ( ) windows of the expert system. figure ( ) some programmable steps of the expert system. no. vol. al-rafidain engineering figure ( )window contains the menu bar and the boundary conditions figure ( ) window displays when the system beginning. no. vol. al-rafidain engineering figure ( ) window used to calculate wave speed. no. vol. al-rafidain engineering figure ( ) window to select pump station type. figure ( ) window for analysis horizontal pump power failure. no. vol. al-rafidain engineering figure ( ) window for analysis vertical pump power failure. figure ( ) window for analysis of surge tank in network. no. vol. al-rafidain engineering figure ( ) window for analysis air chamber in network. figure ( ) window for analysis surge valve in network. no. vol. al-rafidain engineering figure ( ) window for analysis dead end in network. figure ( ) window for analysis networks (branched pipes). no. vol. al-rafidain engineering figure ( ) window for analysis network has valve at downstream. figure ( ) window for analysis network has jet at downstream. insight sfi research centre for data analytics search insight menu about who we are what we do our structure people work with us senior leadership principal investigators funded investigators research and operations research application domains demonstrators research challenges core scientific expertise publications projects european funded projects business masterclasses business team public engagement epe committee citizen science news latest news media queries newsletter spotlight on research contact search search for: close search close menu aboutshow sub menu who we are what we do our structure peopleshow sub menu work with us senior leadership principal investigators funded investigators research and operations researchshow sub menu application domains demonstrators research challenges core scientific expertise publications projects european funded projects businessshow sub menu masterclasses business team public engagementshow sub menu epe committee citizen science newsshow sub menu latest news media queries newsletter spotlight on research contact insight impact ep. happy poddy’s day! listen to the final episode of the insight podcast read on insight careers insight is hiring! check out our jobs page read on work with us insight impact professor barry o’sullivan wins the prestigious nerode prize read on insight business visit our business team and read industry case studies insight research research informed by our vision: enabling citizens. smarter societies. insight people find an expert. work with us. public engagement school visits, events, citizen science, public education and more latest news ‘people believe vr is all about games!’ lukasz porwol talks to silicon repu... / / insight partners with global leader avaya on next gen ai / / prof alan smeaton talks deepfakes with newstalk’s pat kenny / / ep. happy poddy’s day! listen to the final episode of the insight podcast / / empowering citizens smarter societies find out more insight in numbers + researchers € + million in funding + industry partners research institutions stay up to date with our newsletter receive all the latest news & insights* leave this field empty if you're human: privacy statement copyright statement data protection notice tai bgo.eps memetic search for the quadratic assignment problem una benlica, jin-kao haob,∗ achords, university of stirling, stirling fk la, united kingdom bleria, université d’angers, bd lavoisier, angers, cedex , france accepted to expert systems with applications, august abstract the quadratic assignment problem (qap) is one of the most studied np- hard problems with various practical applications. in this work, we propose a powerful population-based memetic algorithm (called bma) for qap. bma in- tegrates an effective local optimization algorithm called breakout local search (bls) within the evolutionary computing framework which itself is based on a uniform crossover, a fitness-based pool updating strategy and an adaptive mu- tation procedure. extensive computational studies on the set of well-known benchmark instances from the qaplib revealed that the proposed algorithm is able to attain the best-known results for instances and thus competes very favorably with the current most effective qap approaches. a study of the search landscape and crossover operators is also proposed to shed light on the behavior of the algorithm. keywords: memetic algorithm; local search; landscape analysis; quadratic assignment; combinatorial optimization. . introduction the quadratic assignment problem (qap) is a classic np-hard combinatorial optimization problem with a number of applications (cheng, ho, & kwan, ; garey & johnson, ; li, xu, jin, & wang, ; miao, cai, & xu, ; nikolić & teodorović, ; pardalos, rendl, & wolkowicz, ). qap is to determine a minimal cost assignment of n facilities to n locations, given a flow aij from facility i to facility j for all i, j ∈ { , ..., n} and a distance bqp between locations q and p for all q, p ∈ { , ..., n}. let Π denote the set of the permutation functions π : { , ..., n} → { , ..., n}, then qap can mathematically be formulated as follows: minπ∈Πf(π) = n∑ i= n∑ j= aijbπiπj ( ) ∗corresponding author. email addresses: ube@cs.stir.ac.uk (una benlic), hao@info.univ-angers.fr (jin-kao hao) preprint submitted to elsevier august , where a and b are the flow and distance matrices respectively, and π ∈ Π is a solution where πi represents the location chosen for facility i. the problem objective is then to find a permutation π∗ in Π that minimizes the sum of the products of the flow and distance matrices, i.e., f(π∗) ≤ f(π), ∀π ∈ Π. besides the facility location problem, qap is notable for its ability to for- mulate a number of other practical problems such as backboard wiring in elec- tronics, design of typewriter keyboards, campus planning, analysis of chemical reactions for organic compounds, balancing turbine runners, and many others. qap can equally formulate some classic combinatorial optimization problems such as the traveling salesman, maximum clique and graph partitioning prob- lems. reviews on some significant applications of qap can be found in burkard ( ); duman & or ( ) and pardalos, rendl, & wolkowicz ( ), while many solution methods are reviewed in anstreicher ( ). qap is among the most studied and the hardest combinatorial optimization problems. in fact, from a theoretical point of view, qap is np-hard (garey & johnson, ). consequently, no exact algorithm is expected to solve the problem in a polynomial time and even small instances may require considerable computation time. this hardness is confirmed in practice since the existing ex- act algorithms can solve to optimality only small instances from the qap bench- mark library with up to locations. even approximation of the problem with a guaranteed performance is known to be very hard (hassin, levin, & sviridenko, ). for these reasons, heuristic and metaheuristic methods constitute a nat- ural and useful approach for tackling this problem (blum, puchinger, raidl, & roli, ). such algorithms aim to provide satisfactory sub-optimal solutions in acceptable computing time, but with no theoretically provable guarantee that the attained solutions are the optimal ones. performance of these heuristic al- gorithms is typically assessed using a set of benchmark instances. among the numerous heuristic algorithms reported for qap in the literature, local search methods are very popular approaches, including simulated annealing (wilhelm & ward, ), tabu search (battiti & tecchiolli, ; james, rego, & glover, a,b; misevicius & kilda, ; skorin-kapov, ; taillard, ), and iterated local search (benlic & hao, c; stützle, ). population-based approaches constitute another class of popular tools for finding high quality near-optimal solutions for qap (ahuja, orlin, & tiwari, ; drezner, , ; fleurent & ferland, ; merz & freisleben, ; misevicius, ; stützle, ). in this work, we are interested in solving qap with heuristic algorithms. we introduce a powerful memetic algorithm (bma) for qap which combines an effective local search algorithm (bls - breakout local search), a crossover operator, a pool updating strategy, and an adaptive mutation mechanism. bls is a general local search method which has shown very good results for several np-hard problems including maximum clique (benlic & hao, a), maximum cut (benlic & hao, b), qap (benlic & hao, c) and vertex separator (benlic & hao, d). its basic idea is to use a descent procedure to discover local optima and employ dedicated perturbations to continually move from one attractor to another in the search space. in this paper, we integrate bls into the memetic framework, thus extending our work of benlic & hao ( c). as we show in this paper, the proposed memetic algorithm bma exhibits an excellent performance on the whole set of well-known qap benchmark instances. indeed, bma attains the best-known solution for all the instances except for only two cases. furthermore, it outperforms its local search procedure bls which confirms the usefulness of the memetic framework. in order to gain some insight into the functioning of the proposed memetic algorithm, we perform a landscape analysis and justify the choice for the used crossover operator (see section ). the paper is organized as follows. in the next section, we briefly review the most effective qap approaches and highlight the contributions of this work. in section , we present our proposed memetic algorithm (bma) and detail its main components. moreover, we highlight the differences and similarities between bma and the reviewed state-of-art approaches. computational results and comparisons with the top-performing qap algorithms are presented in section . section provides a landscape study that we use to justify the choices for the crossover operator of our memetic algorithm, and additionally shows a comparison of several crossover operators. conclusions are provided in the last section. . state-of-art approaches for qap and main contributions in this section, we provide a literature review of the most popular heuristic approaches for qap, including four population-based algorithms (drezner, ; merz & freisleben, ; misevicius, ; stützle, ) and two local search algorithms (james, rego, & glover, a; misevicius & kilda, ), followed by a summary of the main contributions of this study. in section . , we discuss in more detail the relationships between these approaches and the proposed bma. among the reviewed approaches, four algorithms from drezner ( ); james, rego, & glover ( a); misevicius ( ) and misevicius & kilda ( ) report the best results on some particular class of qap benchmark instances. for this reason, these four algorithms will be used as reference algorithms for our comparative study. nevertheless, it is important to mention that none of the existing qap approaches can be considered as the most effective for all the different types of qap instances, due to significant differences in structure of these instance (see section ). a popular memetic approach for qap (ma-qap) is introduced in merz & freisleben ( ) which incorporates the -opt local search procedure and an adaptation of the standard uniform crossover ux that does not perform any implicit mutation. the selection for reproduction is performed on a purely random basis, while the selection for survival is achieved by choosing the best individuals from the pool of parents and children. to overcome premature convergence, the restart technique proposed in eshelman ( ) is employed. during the run, it is checked whether the average distance of the population has dropped below a certain threshold or whether the average fitness of the population did not change after a certain number of consecutive generations. if one of these conditions holds, the whole population is mutated except the best individual, and each mutated individual is improved by the -opt local search to obtain a local optimum. afterwards, the algorithm proceeds with the usual recombination process. we mention this work since it is one of the first memetic algorithms applied to qap and achieved remarkable results at the time it was published. the improved hybrid genetic algorithm (ihga) (misevicius, ) incor- porates a robust local improvement procedure as well as an effective restart mechanism based on shift mutations. the author slightly improved the clas- sic scheme of a uniform like crossover (ulx) to get a new optimized crossover (ox). the optimized crossover is a crossover that (a) is ulx and (b) produces a child that has the best fitness value among the children created by m runs of ulx. the offspring is then improved with a local search mechanism, based on the swap neighborhood, which contains a tabu search procedure and a so- lution reconstruction procedure. the reconstruction is achieved by performing a number µ of random swaps, where µ is varied according to the instance size. once convergence of the algorithm is observed, all the individuals but the best undergo the shift mutation (sm), which simply consists in shifting all the items in a wrap-around fashion by a predefined number of positions. ihga is one of the best-performing algorithms for the unstructured instances and real-life like instances and is used as one of the references in our comparative study. drezner ( ) shows extensive computational experiments on qap using various variants of a hybrid genetic algorithm. the author compared the modi- fied robust tabu search (mrt) and the simple tabu search as local optimization algorithms combined with a crossover operator. moreover, different parent selec- tion (distance modification, gender modification) and pool updating strategies were tested. the best version of the memetic algorithm is mrt which inte- grates the modified robust tabu search mrt for offspring improvement. mrt is identical to the robust tabu search (rts) (taillard, ) except that the tabu tenure is generated in [ . n, . n] rather than in [ . n, . n]. mrt is the best-performing memetic algorithm for the grid-based instances and is used as another reference algorithm in our comparative study. population-based iterated local search (pils) (stützle, ) is another highly effective algorithm. the underlying iterated local search (ils) algorithm starts from a random assignment, and applies a first-improvement local search procedure based on the -opt neighborhood. to speed up the search process, the algorithm uses the don’t look bit strategy, previously proposed to accelerate local search algorithms for tsp. once a local optimum is reached, ils applies a perturbation that consists of exchanging k randomly chosen items, where k is varied between kmin and kmax. stützle extends the described ils to a population-based algorithm where no interaction between solutions takes place, and each single solution is improved by the standard ils. in the proposed pils, the population consists of µ solutions and in each iteration λ new solutions are generated. a selection strategy, based both on quality and distance between solutions, is then employed to form a new population of µ solutions from the set of µ + λ solutions. cooperative parallel tabu search algorithm (cpts) (james, rego, & glover, a) executes in parallel several tabu search (ts) operators on multiple pro- cessors. the ts operator is a modified version of taillard’s rts (taillard, ) obtained by changing the stopping criterion and the tabu tenure parameters for each processor participating in the algorithm. in order to accomplish coopera- tion between ts processes, cpts maintains a global reference set which uses information exchange to promote both intensification and diversification in a parallel environment. cpts globally obtains excellent results on the whole set of qap instances and is used as another reference algorithm in our comparative study. iterated tabu search (its) by misevicius & kilda ( ) follows the general scheme of the iterated local search metaheuristic. it employs a traditional tabu search to reach local optima and triggers a perturbation (reconstruction) phase in order to escape the attained local optimum. the “ruined” solution becomes the new starting point for the basic ts procedure. its uses a perturbation mechanism which varies adaptively the number of random perturbation moves in some interval. its obtains excellent results on the unstructured instances and real-life like instances and is used as another reference in our comparison. compared to the existing studies on qap, this work has the following main contributions: first, the proposed bma algorithm is based on a new local optimization pro- cedure (i.e., bls) which adopts an adaptive perturbation mechanism to better escape local optima. the computational study discloses that this algorithm performs very well on the set of very popular qap benchmark instances by attaining the best-known result in cases. this work thus confirms the usefulness of the memetic framework for qap. second, we provide an empirical justification to explain why the uniform crossover is the best operator for qap in comparison with the other crossover operators studied in the literature. third, ideas of the proposed algorithm could help in designing effective heuristics for other related permutation problems and applications such as those mentioned in the introductory section. . an effective memetic algorithm for qap the term memetic algorithm (ma) is employed to designate a general heuris- tic approach which typically combines local optimization with a population- based paradigm (moscato, ; moscato & cotta, ). the purpose of such a combination is to take advantages of both crossover that discovers unexplored promising regions of the search space, and local optimization that finds good solutions by concentrating the search around these regions. since memetic al- gorithm is a problem-independent framework, it needs to be properly adapted to the specific problem at hand to ensure the best performance (hao, ; krasnogor & smith, ). in particular, local optimization operator and re- combination operator are the two key components to consider. finally, as for any population-based method, a healthy diversity of population must be main- tained to avoid premature convergence. previous studies show that memetic algorithms are able to achieve excellent performances for a number of optimiza- tion problems (hart, krasnogor, & smith, ; moscato & cotta, ; neri, cotta, & moscato, ). given an initial population which consists of locally optimal solutions, a memetic approach generates new solutions by applying crossover and/or mu- tation, followed by a phase of local search to improve each offspring solution. the right choice for a crossover operator depends on problem structure and landscape properties (in section , we show a study on the relationship between the performance of a crossover operator and the structural properties of a given problem). moreover, the success of a memetic approach is conditioned by the ef- fectiveness of the local search procedure. while the main role of the crossover is to discover unexplored promising regions of the search space (i.e., exploration), local search basically aims to find good solutions by concentrating the search around these regions (i.e., exploitation). the proposed memetic algorithm for qap (bma) employs the uniform crossover operator (section . ) (in section . , we justify why we chose this particular crossover), a breakout local search (bls) procedure (section . ), a fitness-based population replacement strategy (section . ), and an adaptive mutation mechanism (section . ). each offspring solution, generated with the uniform crossover, is improved with the bls procedure. our memetic approach then applies a pool updating strategy to possibly replace the worst individual from the population with the improved offspring solution. to avoid premature convergence, bma triggers an adaptive mutation mechanism to the entire popu- lation if the best solution found during the search has not been improved during a fixed number of generations. this displaces the search to distant regions each time a search stagnation is detected. the general architecture of our memetic approach is described in algorithm . the main components are detailed in the following sections. . . parents selection and recombination to determine a subset p ⊂ p of |p| parent individuals, we employ the tour- nament selection strategy. let λ be the size of the tournament pool. we select each individual πi ∈ p in the following way: randomly choose λ individuals from p ; among the λ chosen individuals, place the best one into p if it is not already present in p. the time complexity of this operation is o(|p |). an advantage of the tournament selection is that the selection pressure can easily be adjusted by changing the size of the tournament pool λ. the larger the tournament pool is the less is the chance to select weaker individuals. an experimental evaluation of our qap approach has revealed that putting a higher pressure on better individuals gives better results than using a random selection. for the crossover process, we employ a standard uniform operator (ux) that recombines two parent individuals from subset p. elements of the parents are scanned from left to right and each element in the offspring keeps with equal probability the value j ∈ [ ...n] found in either of the two parents, under the algorithm general scheme of the proposed bma algorithm : initialize the number of bls iterations for short and long runs ts and tl respectively, the minimum mutation degree µmin and the increment m of mutation degree : randomly generate initial population p : p ← bls(p, ts) /* improve each individual with ts iterations of bls, sect. . */ : π best ← bestindividual(p) /* initialize the best individual */ : µ ← µmin /* initialize the current mutation degree */ : for i := to number of generations φ do : select a subset of parent individuals p from p /* sect. . : π ← crossover(p) /* generate an offspring, sect. . */ : π ← bls(π , tl) /* improve offspring π with long bls run, sect. . */ : p ← replacementstrategy(p, π ) /* sect. . */ : if (πbest not improved after ν generations) then : p ← mutate(p, µ) /* sect. . */ : p ← bls(p, ts) : update the best solution πbest if necessary : µ ← µ + m /* increase mutation degree */ : end if : if (f(πbest) > f(π ) or µ > n) then : µ ← µmin /* reset mutation degree to default */ : end if : if (f(πbest) > f(π )) then : π best ← π : end if : end for constraint that j has not been assigned before to any element in the offspring. any unassigned element is given a random value j such that j does not appear twice in the offspring. an example of this recombination process is illustrated in figure . the complexity of this crossover is o(n). since the uniform crossover permits great flexibility, different variations of its basic procedure have been proposed and applied to qap (merz & freisleben, ; misevicius, ). despite its simplicity, the ux has shown to provide very good results on different combinatorial optimization problems including qap. a comparison of the simple ux with several other crossover operators used for qap is reported in section . . . . breakout local search (bls) the local optimization procedure has a significant impact to the overall performance of a memetic algorithm. in our case, we adopt an existing local search algorithm bls-qap (benlic & hao, c) (call it bls for simplicity). as most local search algorithms for qap, bls employs the swap operator which consists in exchanging the values from two different positions in π, i.e., permuting the locations of two facilities. it uses the best improvement descent procedure to exploit the whole swap neighborhood n(π), which is evaluated in figure : an example of the uniform crossover (ux) o(n ) time thanks to an effective neighborhood evaluation strategy proposed in taillard ( ). once a local optimum is returned by the best improvement descent pro- cedure, bls triggers a multi-type diversification mechanism which adaptively determines the type t of perturbation moves and the number l of perturbations (called jump magnitude) by considering some information related to the search state. this mechanism combines two complementary types of perturbation: a guided perturbation (using a tabu list) and a random perturbation. the tabu- based perturbation uses a selection rule that favors swap moves that minimize the cost degradation, under the constraint that the move has not been applied during the last γ iterations (where γ is the tabu tenure that takes a random value from a given range), while the random perturbation performs moves se- lected uniformly at random. in order to insure a good balance between an intensified and a diversified search, bls-qap alternates probabilistically be- tween the two types of perturbations. the probability of applying a particular perturbation is determined dynamically depending on the current number ω of consecutive non-improving attractors visited (see benlic & hao ( c) for more details). we limit the probability of applying the tabu-based perturbation over the random perturbation to take values not less than q. to determine the jump magnitude l for the following perturbation phase, the proposed algorithm uses a basic strategy which increments l if the local search procedure returned to the immediate previous local optimum, and otherwise resets l to its initial value l . once the type t and the number l of perturbations are determined, we apply accordingly l moves of type t to the current solution. the resulting solution is used by the next round of the best improvement descent procedure as its new starting point. once an offspring solution is created in the crossover phase (line of alg. ), it is improved with tl iterations of the bls procedure (line of alg. ). as previously mentioned, the complexity of each iteration of the bls descent procedure is o(n ). the number of the descent iterations performed in each iteration of bls depends on the size sbasin of the basin of attraction of a local optimum. this, in turn, depends on the type of perturbation used for the previous perturbation phase. more precisely, after a phase of the directed (tabu- based) perturbation, the descent-based local search requires, on average, less steps to attain a local optimum than after the random perturbation. however, the computational complexity of one iteration of the tabu-based perturbation is o(n ), while the complexity of the random perturbation is o( ). therefore, the time complexity of one bls iteration is bounded within o(sbasin·n )+o(l·n ). . . pool updating strategy for each offspring solution π created by the crossover operator and im- proved with the bls procedure (section . ), we decide whether π should be inserted into the population pool (lines - of alg. ). to base this decision, our algorithm uses a classic replacement strategy which inserts π into p if there is no individual in p identical to π (i.e., ∀πi ∈ p, πi = π ), and if the fitness of π is better than the fitness of the worst individual πworst from p (i.e., f(π ) < f(πworst)). if this condition holds, p is updated by replacing πworst with π . to determine whether an identical solution to π is already present in p , we compute the hamming distance between π and each πi ∈ p . the ham- ming distance is defined as the number of positions at which the corresponding elements are different. if two solutions are identical, the corresponding ham- ming distance is . the overall time complexity of this pool update strategy is thus o(|p | · n). one potential risk of this replacement strategy is the premature loss of popu- lation diversity, since offspring is being inserted into the population regardless of its distance to other individuals in the population. to avoid this problem, more sophisticated pool updating strategies have been proposed in the literature that maintain population diversity by considering the similarity (distance) between the offspring and other individuals from the population (see for instance benlic & hao ( ); lü & hao ( ); porumbel, hao, & kuntz ( )). in our case, the diversity is maintained by an adaptive mutation mechanism described in section . . as a result, the proposed memetic approach is not really sensitive to the pool replacement strategy employed. . . adaptive mutation procedure as soon as a search stagnation is detected, i.e., the best solution πbest found during the search has not been updated for a certain number of generations, our bma algorithm mutates the entire population with an adaptive mutation procedure that adjusts the diversification strength µ ∈ [µmin, n] depending on the previous search progress (see lines - , algorithm ). at the beginning, µ is set to µmin and is gradually augmented by an increment m if a solution better than πbest has not been attained during ν generations. once a solution better than πbest is obtained, or µ has reached the maximum possible value n (n being the problem size), µ is reset to µmin (see lines - , algorithm ). figure : an example of mutation with exchanges for µ = the mutation procedure of our memetic algorithm, which was previously used in merz & freisleben ( ) for qap, exchanges a sequence of locations in the solution π to create an offspring that has a distance of µ to its parent π. the distance measure used for this mutation operator is the well-known hamming distance. to ensure that the distance between the offspring and its parent is µ, in each step, the second selected location is swapped again in the subsequent step, such that the resulting distance is equal to one plus the number of swaps. an example of the mutation procedure is illustrated in figure . the complexity of the mutation procedure is o(|p | · µ). . . discussions in this section, we discuss similarities and differences between bma and the previously mentioned state-of-art qap approaches, in particular those based on the memetic framework. in section , we reviewed three popular memetic qap algorithms, ma-qap (merz & freisleben, ), ihga (misevicius, ), and mrt (drezner, ), the two latter ones being among the best performing algorithms cur- rently available in the literature for this problem. the genetic operators em- ployed by bma are quite similar to those of ma-qap and ihga. indeed, these approaches create offspring solution by means of different adaptations of the classic uniform crossover operator which provides surprisingly good results for the qap instances. one of the contributions of our work is an explanation for the efficiency of these random crossovers obtained by analyzing the structures of the qap instances (see section ). moreover, as ma-qap and ihga, bma applies a mutation mechanism to the entire population in order to avoid premature convergence. however, unlike the previously used mutation strategies, the mutation mechanism of bma adaptively adjusts the diversification strength depending on the previous search progress. yet, the most important difference between bma and the three reference memetic algorithms is the local search procedure. as stated earlier, bls, em- ployed by bma for offspring improvement, constitutes the key component of bma. it combines the steepest descent with a dedicated and adaptive diver- sification mechanism. on the other hand, the local search phase performed by ma-qap is the basic steepest descent algorithm, while ihga and mrt improve solutions by means of a simple tabu search procedure. pils (stützle, ) is a population-based approach which uses an adap- tation of the iterated local search framework to improve each newly generated solution. although the perturbation mechanism of pils determines the number of perturbation moves in an adaptive way, it is only based on random moves which constitutes the main difference with our bls procedure. two local search algorithms, cpts (james, rego, & glover, a) and its (misevicius & kilda, ) (see section ), are among the top performing qap methods. even though bls’s directed perturbation is based on the use of a tabu list, cpts is not very much related to bls since it consists of a parallel execution of several tabu search (ts) operators on multiple processors. the proposed bls is more related to its since both algorithms are variants of the iterated local search method. nevertheless, there are two major differences between its and bls. firstly, bls does not consider the tabu list during its local search (descent) phases, unlike its which constraints each iteration of its local search phase with a tabu list. as such, bls and its explore different trajectories during their respective search, leading to different local optima. in fact, one of the keys to the effectiveness of bls on qap is that it completely excludes diversification during local search, unlike tabu search for which the intensification and diversification are always intertwined. secondly, unlike its which applies solely random moves to perturb the current local optimum, bls adaptively choses between two types of moves (random and directed) according to the search status, leading to variable levels of diversification. finally, we mention the approach proposed by kelly, laguna, & glover ( ). in their work, the authors present a method which also starts with the steepest descent to find local optima, and then applies a diversification strategy to incrementally restrict the set of allowed swap moves in order to escape local optima. the diversification procedure of this approach exploits the search history as well, but unlike bls, performs perturbations in a purely deterministic way. it consists of imposing a “maximum tabu tenure” to moves performed during each descent of local search, and applying the best move among the non prohibited ones to perturb a local optimum. . experimental evaluation . . benchmark instances it is worthy to recall that given the very nature of heuristic algorithms, it is a common practice to evaluate the performance of a new algorithm by testing it on well-established benchmark instances of the problem and comparing its results table : settings of important parameters. para. description value |p | population size ts number of iterations for short bls run tl number of iterations for long bls run µmin minimal mutation degree . n m increment of mutation degree . n λ the size of the tournament pool ν number of generations without improvement before mu- tation |p | l initial jump magnitude of bls . n (t. i & ii), . n (t. iii & iv) γ tabu tenure for directed perturb. with bls random[ . n, . n] q smallest probability for applying bls directed perturba- tion . with those of the state-of-art methods. in our case, we evaluate the performance of our bma algorithm on the set of instances from the qaplib , whose size n varies from to and is indicated in the instance name. these instances are very popular and largely used in the literature. they are typically classified into four types covering various real applications and random problems: type i. real-life instances obtained from practical applications of qap; type ii. unstructured, randomly generated instances for which the distance and flow matrices are randomly generated based on a uniform dis- tribution; type iii. randomly generated instances with structure that is similar to that of real-life instances; type iv. instances in which distances are based on the manhattan distance on a grid. among the instances from the qaplib, we focus on the set of selected instances. the remaining instances (including all the real-life instance of type i) are omitted from our experimental comparisons since bls and bma (and many other state-of-art qap methods) can solve them to optimality in every single trial within a very short computation time (often less than a second). however, to be exhaustive, we include the results of our bma algorithm for these instances in the appendix (table ). . . experimental protocol the proposed memetic algorithm bma is programmed in c++, and com- piled with gnu g++ on a xeon e with . ghz and gb. like other qap heuristic algorithms, bma requires a number of parameters to be tuned (see urlhttp://www.seas.upenn.edu/qaplib/inst.html the code of our memetic algorithm, used to obtained the reported results, will be made available online at http://www.info.univ-angers.fr/pub/hao/bma.html table ), most of which are related to its bls local optimizer and adopt the values used in benlic & hao ( c). according to the parameter analyses per- formed in benlic & hao ( b,c), the most relevant bls parameters are the initial jump magnitude l and the smallest probability for applying directed over random perturbation q. moreover, as discussed in benlic & hao ( c), the optimal setting of these parameters depends on the landscape properties of the qap instance at hand and may vary for different instances. additionally, following the existing ma literature on discrete optimization (benlic & hao, ; bontouxa, artigues, & feillet, ; dorne & hao, ; hao, ; lü & hao, ; merz & freisleben, ), bma maintains a population of fairly limited size. this, as well as the other parameters related to the genetic algo- rithm components, is determined with a preliminary experiment. even though it is possible to find a configuration of parameters better than the one used in this paper, the computational experiments reported in this section show that the adopted setting performs globally well on the tested benchmark instances. more generally, since the source code of our bma algorithm will be made avail- able online, the potential user can possibly apply any preferred method to tune these parameters. according to the practice in the qap literature, the stopping condition is the elapsed time. in our case, we set the maximum time limit to hours for all the instances, except for the largest instance tho for which we set the maximum time allowed to hours. notice that some reference qap algo- rithms (drezner, ; james, rego, & glover, a) use an even higher time limit for tho to reach the reported result. as shown below, the best-known solutions are very often attained long before these time limits. furthermore, we provide computational results of our bma algorithm with the time limit reduced to minutes. the reported results for bma and bls are obtained over independent executions. we focus primarily on the comparisons in terms of the solution quality with respect to the best-known results (bkr) reported in the literature, which were obtained with different qap algorithms under various conditions. for indicative purposes, we also compare the results produced by our memetic approach and those obtained with the state-of-art qap approaches which are the current best performing methods: . cooperative parallel tabu search (cpts) algorithm by james, rego, & glover ( a). the reported results are obtained using ten ( . ghz) intel intanium processors; . iterated tabu search (its) by misevicius & kilda ( ). the results are obtained using a mhz pentium computer; . improved hybrid genetic algorithm (ihga) by misevicius ( ). the reported results are obtained on a x family processor; . a variant of a hybrid genetic tabu search algorithm (mrt ) by drezner ( ). the reported results are obtained on a pentium iv . ghz com- puter. an exhaustive comparative analysis with the reference approaches is not a straightforward task because of the differences in computing hardware, re- sult reporting methodology, termination criterion, etc. this comparison is thus presented only for indicative purposes. nevertheless, this experiment provides interesting indications on the performance of the proposed algorithm relative to the state-of-art algorithms. in addition, we evaluate the contribution of the proposed memetic approach by comparing (under the same conditions) bma with its bls procedure (benlic & hao, c) which itself shows to be highly effective on the tested qap instances. this allows us to highlight the usefulness of the memetic framework. according to the literature (drezner, ; misevicius, ; stützle, ), we employ several criteria for evaluation and comparison of our memetic algo- rithm with other qap approaches. . the number of instances for which an optimal or best-known solution is reached within a reasonable computing time. this constitutes an indicator on the effectiveness of an algorithm in terms of the solution quality. . the success rate of reaching an optimal or best-known solution which provides information about the robustness of an algorithm. . the percentage deviation δ̄avg of the average solution from the published best-known result over a certain number of runs. the percentage deviation between solutions is computed as δ̄ = (z − z̄)/z̄[%], where z is the average result over a given number of runs and z̄ the best-known objective value. this indicator provides additional information about the robustness of an algorithm. beside the aforementioned criteria, we also mention the amount of compu- tational time required by each approach to reach the reported results. this pro- vides an indication about the computational efficiency of an algorithm. more- over, table in the appendix provides a detailed summary of the computational results for selected qaplib instances obtained by bma within a time limit of hours ( hours for instance tho ) and minutes respectively. . . computational results and comparisons table reports comparative results with bls (section . ), cpts (james, rego, & glover, a), its (misevicius & kilda, ) and ihga (misevicius, ), for unstructured instances (type ii) and real-life like instances (type iii). since mrt (drezner, ) does not report results on these instances, it is excluded from this comparison (in fact, mrt only reports results for instances of type iv). the second column ‘bks’ shows for each instance the best-known objective value ever reported in the literature. for each algorithm, column δ̄avg indicates the percentage deviation between an average solution, obtained with the given approach over independent trials, and the best-known solution. the success rate for reaching the best-known solution over trials is given in table : comparative results between the proposed memetic algorithm (bma), bls (benlic & hao, c), and three best performing qap approaches on unstructured instances (type ii) and real-life like instances (type iii): cpts (james, rego, & glover, a), its (misevicius & kilda, ) and ihga (misevicius, ). the success rate of reaching the best-known result over executions is indicated between parentheses. computing times are given in minutes for indicative purposes. problem bks bma bls cpts its ihga % δ̄avg t(m) % δ̄avg t(m) % δ̄avg t(m) % δ̄avg t(m) % δ̄avg t(m) random instances (type ii) tai a . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . tai a . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . tai a . ( ) . . ( ) . . ( ) . . ( ) . . ( ) tai a . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . tai a . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . average . . . . . . . . . . real-life like instances (type iii) tai b . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . tai b . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . tai b . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . tai b . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . tai b . ( ) . . ( ) . . ( ) . . ( ) . . ( ) . average . . . . . . . . . . table : comparative results between the proposed memetic algorithm (bma), bls (benlic & hao, c), cpts (james, rego, & glover, a) and mrt (drezner, ) on grid-based (type iv) instances. the success rate of reaching the best-known result over executions is indicated between parentheses. computing times are given in minutes for indicative purposes. problem bks bma bls cpts mrt % δ̄avg t(m) % δ̄avg t(m) % δ̄avg t(m) % δ̄avg t(m) sko . ( ) . . ( ) . . ( ) . . ( ) . sko . ( ) . . ( ) . . ( ) . . ( ) . sko . ( ) . . ( ) . . ( ) . . ( ) . sko a . ( ) . . ( ) . . ( ) . . ( ) . sko b . ( ) . . ( ) . . ( ) . . ( ) . sko c . ( ) . . ( ) . . ( ) . . ( ) . sko d . ( ) . . ( ) . . ( ) . . ( ) . sko e . ( ) . . ( ) . . ( ) . . ( ) . sko f . ( ) . . ( ) . . ( ) . . ( ) . wil . ( ) . . ( ) . . ( ) . . ( ) . tho . ( ) . . ( ) . . ( ) . . ( ) . average . . . . . . . . parentheses next to the value of δ̄avg. the cpu time (in minutes) is only given for indicative purposes. from table , we can make the following observations. for the unstructured instances (type ii), bma finds the best-known solution for out of the instances, with an average deviation δ̄avg of . over the instances. for three hard instances tai a, tai a and tai a, it reaches the best-known solution in % of the trials. compared to its local search procedure bls, bma statistically outperforms (with p-value< . ) bls for two hard instances, tai a and tai a, and reduces the average percentage deviation from . to . over the instances. the current most effective approach for instances of type ii is probably its (misevicius & kilda, ), which attains the best-known result for out of the instances with an average deviation δ̄avg of . . one notices that its shows worse results than both bma and bls, but it requires shorter computing time. to verify the performance of bma under short computing budgets, we show in appendix (table ) the results of bma within a significantly shorter cutoff limit ( minutes). from table , we observe that bma does remain very competitive with its, with an average deviation δ̄avg of . over the instances. finally, the two other reference algorithms cpts and ihga achieve results which are slightly worse than bma, bls and its. for the real-life like instances (type iii), bma is able to attain the best- known solution for all the instances in every single trial, except for the largest instance tai b where the success rate is %. it reports an average deviation δ̄avg of . over the instances. unlike bma, bls is unable to reach the best-known solution for tai b with the given computing conditions. when the running time of bma is greatly reduced, it is still able to attain the best- known solution for all the five instances but with an average deviation δ̄avg of . (see table ). two reference approaches, its (misevicius & kilda, ) and ihga (misevicius, ), are also able to attain the best-known result for all the real-life like instances with an average deviation δ̄avg of . and . respectively. we now turn our attention to the instances with grid distances (type iv, table ). among the reference methods, only cpts (james, rego, & glover, a) and mrt (drezner, ) report results for these instances. in table , we show the results of our bma and bls algorithms together with those of cpts and mrt for these type iv instances. from table , we observe that bma is able to reach the best-known result for all the instances of this type. for out of instances, it has a success rate of %. for the two remaining instances (sko d and tho ), the success rate is % and % respectively. bls also attains the best-known result for the hardest instance tho with a success rate of %. mrt has a slightly better success rate than bma on instance sko d ( / v.s. / ) and reports a slightly worse success rate than bma on sko f ( / v.s. / ). on the other instances, bma and mrt have the same success rates with a very slight advantage for mrt in terms of the average gap over all the instances. on the other hand, cpts performs globally well except for the hardest instance tho for which cpts fails to attain the best-known solution. we conclude that bma competes favorably with mrt and cpts on the type iv instances. . analysis and discussion the performance of a memetic algorithm may be influenced by the character- istics of the search landscape like the average distance between local optima and the relative distance of local optima to the nearest global optimum. analyses of correlation between solution fitness and distance to global optimum have shown to be particularly useful for a better understanding of algorithm behavior and for designing suitable operators for a more effective search. landscape studies have been previously reported for qap in merz & freisleben ( ) and stützle ( ) and for other well-known problems like the tsp problem (boese, ) and the flow-shop scheduling problem (reeves, ). in this section, we report a similar analysis, based on solutions sampled by our bls procedure which are probably quite different from the solutions used in the previous studies on qap. we perform the landscape analysis on qaplib instances, based on a set of distinct solutions obtained after independent runs of our bls approach. the number of iterations per run is set to . as the distance between solutions, we calculate the number of facilities that are allocated to distinct locations in two solutions π and π′, i.e., d(π, π′) = |{i|πi = π ′ i}|. since global optima for the analyzed instances are not known, we use instead the best-known local optima to compute fitness-distance correlation and refer to them as global optima. . . landscape analysis - fdc and distribution of local optima the fitness distance correlation (fdc) coefficient ρ (jones & forrest, ) captures the correlation between the solution fitness and its distance to the nearest global optimum (or best-known solution if global optimum is not avail- able). for a minimization problem, if the fitness of a solution decreases with the decrease of distance from the optimum, then it should be easy to reach the target optimum for an algorithm that concentrates around the best candidate solutions found so far, since there is a “path” to the optimum via solutions with decreasing (better) fitness. a value of ρ = indicates perfect correlation between fitness and distance to the optimum. for correlation of ρ = − , the fitness function is completely misleading. fdc can also be visualized with a fd plot, where the same data used for estimating ρ is displayed graphically. . . . . . . . . % d e vi a tio n t o t h e b e st k n o w n s o lu tio n distance to optimum tai a . . . . . . % d e vi a tio n t o t h e b e st k n o w n s o lu tio n distance to optimum sko a . . . % d e vi a tio n t o t h e b e st k n o w n s o lu tio n distance to optimum tai b figure : distances of local optima to the best-known solution based on solutions sampled by the bls algorithm for instances tai a (type ii), sko a (type iii) and tai b (type iv). in column ‘ρ’ of table , we report fdc coefficients for the selected qap instances. for illustrative purpose, fd plots for three instances (tai a, sko a and tai b) are given in figure . as it can be seen from the fdc coefficients in table , there is a clear difference in correlation among instances of different types. for randomly generated instances (type ii), the fdc coefficient ρ is negative except in one case (tai a) where ρ is close to zero. indeed, from the fd plots in figure it is clear that there is no correlation between fitness table : analytical results for qap instances. column ‘#dlo’ indicates the number of distinct local optima over independent runs of bls; columns ‘avg dlo’, ‘avg dgo’ and ‘avg dhq’ report respectively the average distance between local optima, the average distance between local and global optima, and the average distance between the highest quality local optima; column ‘ρ’ shows the value of the fitness-distance correlation coefficient. instance #dlo avg dlo avg dgo avg dhq ρ type tai a . . . - . ii tai a . . . . ii tai a . . . - . ii tai a . . . - . ii tai a . . . - . ii tai b . . . . iii tai b . . . . iii sko . . . . iv sko . . . - . iv sko . . . . iv sko a . . . . iv sko b . . . . iv sko c . . . . iv sko d . . . . iv sko e . . . . iv sko f . . . . iv and distance for the random instance tai a. for grid-based instances (type iv), significant fdc exists except in one case (sko , ρ < ), while for two real-life like instances tai b and tai b the fdc is also low ρ < . but higher than for random instances. the fd plots in figure for instances sko a and tai b also confirm this observation. table additionally reports the average distance between local optima avg dlo, the average distance between local optima and the nearest global optimum avg dgo, and the average distance between the highest quality local optima avg dhq. it can be observed that, regardless of the instance type, the values of avg dlo, avg dgo, and avg dhq are very large, close to the maximal possible value n. this implies that local optima are scattered in the search space. even high quality local optima are distant from each other and share no common structure. these observations will allow us to understand why the uniform crossover operator is appropriate for qap. . . comparison of recombination operators the objective of this section is to justify, based on the landscape analysis provided in section . , the choices for the crossover operator used by our bma algorithm. we compare three versions of our bma algorithm incorporating the three recombination operators of the literature for permutation problems: the standard uniform crossover (ux) described in section . , the block crossover (bx) and the distance preserving crossover (dpx). the block crossover (bx), also called the multi-point crossover, is one of the classic recombination operators. firstly, it chooses randomly a certain number of crossing points. once the points have been chosen, starting from left to right, all the elements of the first parent are simply copied over to the offspring π until the first crossing point is reached. then all the elements of the second parent are copied over to π until the second crossing point is reached. this procedure continues until reaching the last position in π . during the crossover process, we constrain an element to take on a value j ∈ [ ...n] that has not already been assigned to some element in π . the rest of unassigned elements are randomly assigned a value j ∈ [ ...n] such that solution feasibility is maintained. the distance preserving crossover (dpx) has previously been applied to qap in merz & freisleben ( ). the basic idea of dpx is to create an offspring that has the same distance to each of its parents, and that distance is equal to the distance between the parents themselves. elements with the same values in both parents are copied to the offspring. the values of all the other elements change. generally, we can classify recombination operators into two classes, those that try to exploit an existing structure in a problem (e.g., the dpx crossover) and those that perform recombination in a purely random way (e.g., ux and bx crossovers). the former operators usually perform well on the class of problems with exploitable global structure and landscapes with highly correlated local optima (e.g., the traveling salesman problem (bontouxa, artigues, & feillet, ) and the graph partitioning problem (benlic & hao, )). otherwise, these operators become destructive introducing a too strong perturbation. table shows the average percentage deviation from the best-known result δ̄avg for the three versions of our bma over runs on qaplib instances. the stopping condition used is the number of generations φ = , with , bls iterations per generation. we indicate in parentheses the number of times each version of bma reached the best-known result over the executions. moreover, we study the degree of perturbation pstr induced by each crossover, and report in table the average pstr caused by the three operators over the , generations. here, we define the perturbation degree pstr as the minimum distance between the created offspring and one of its parents expressed as a percentage of n. from table , we observe that the best performance is obtained by the bma version integrating the standard ux operator, with an average δ̄avg of . over the instances. the second best performing bma version inte- grates the block crossover (bx) with an average δ̄avg of . . on the other hand, the results indicate that dpx is less effective with an average δ̄avg of . over the instances. we further observe that the difference in performance between operators ux & bx and dpx is particularly notable on unstructured instances (type ii) since local optima are the most uncorrelated for these in- stances. indeed, on the other instances with slightly higher fdc coefficient ρ, this difference is much less obvious. as expected, we observe from table that on random, unstructured in- stances, the amount of perturbation induced with dpx is much stronger than with ux and bx, since dpx is designed to exploit the problem structure that is missing for instances of type ii. indeed, the degree of perturbation pstr with dpx crossover on unstructured instances is almost %. on the other hand, the amount of perturbation caused by dpx is greatly smaller when applied to more structured instances of type iv. table : percentage deviations δ̄avg of the average solution (obtained after runs) from the published best-known result for the three versions of our bma integrating respectively the uniform (ux), the block (bx), and the distance preserving (dpx) crossover. instance ux bx dpx instance ux bx dpx tai a . ( ) . ( ) . ( ) sko a . ( ) . ( ) . ( ) tai a . ( ) . ( ) . ( ) sko b . ( ) . ( ) . ( ) tai a . ( ) . ( ) . ( ) sko c . ( ) . ( ) . ( ) tai a . ( ) . ( ) . ( ) sko d . ( ) . ( ) . ( ) tai a . ( ) . ( ) . ( ) sko e . ( ) . ( ) . ( ) sko . ( ) . ( ) . ( ) sko f . ( ) . ( ) . ( ) sko . ( ) . ( ) . ( ) tai b . ( ) . ( ) . ( ) sko . ( ) . ( ) . ( ) tai b . ( ) . ( ) . ( ) table : the average perturbation degree pstr over generations introduced by three versions of our bma integrating respectively the uniform (ux), the block (bx), and the distance preserving (dpx) crossover. pstr is expressed as a percentage of n. instance ux bx dpx instance ux bx dpx tai a . . . sko a . . . tai a . . . sko b . . . tai a . . . sko c . . . tai a . . . sko d . . . tai a . . . sko e . . . sko . . . sko f . . . sko . . . tai b . . . sko . . . tai b . . . . conclusion in this article, we presented a simple and effective memetic algorithm (bma) for the well-known quadratic assignment problem. bma combines a dedicated local search named breakout local search (bls) with the standard uniform crossover (ux), a simple pool updating strategy and an adaptive mutation mechanism. the bls procedure, which is the key component of bma, alternates between a local search phase (to reach local optima) and a dedicated perturbation phase (to discover new promising regions). the perturbation mechanism of bls dy- namically determines the number of perturbation moves and adaptively chooses between two types of perturbation moves of different intensities depending on the current search state. genetic operators (crossover and mutation) are integrated to further enforce the search capacity of bma. the choice of these operators is based on observa- tions made from a landscape analysis which investigates the structure of qap instances. the analysis revealed a very low fitness-distance correlation between local optima for randomly generated instances, and a medium correlation for instances of other types. these observations provide a basis to explain why the standard uniform crossover generally leads to better results for unstructured instances (type ii) than a crossover that tries to exploit the problem structure. we evaluated the proposed algorithm on the set of benchmark instances from the qaplib. computational results revealed that bma performs very well on these instances. indeed, bma outperforms its local search component (bls) and is able to reach the current best-known solution for instances. in particular, bma performs particularly well on unstructured instances (type ii) which are considered to be the hardest for the existing qap methods. for real-life like instances (type iii) and instances with grid distances (type iv), bma remains competitive with respect to the best performing approaches as well. the computing time needed for our proposed algorithm to reach its best solution varies on average from several seconds to about one hour, except for the largest problem for which hours are needed. when the computing time is limited to minutes, bma is still able to attain the best known result for out of the benchmark instances. acknowledgment we are grateful to the anonymous referees for valuable suggestions and com- ments which helped us improve the paper. this work was carried out when the first author was with the leria, université d’angers (france) and was partially supported by the region of ‘pays de la loire’ (france) within the radapop and ligero projects. references ahuja, r.k., orlin j.b., & tiwari, a. ( ). a greedy genetic algorithm for the quadratic assignment problem. computers & operations research, ( ), – . anstreicher, k. m. ( ). recent advances in the solution of quadratic assign- ment problems. mathematical programming, series b , – . battiti, r., & tecchiolli, g. ( ). the reactive tabu search. orsa journal of computing, ( ), – . benlic, u., & hao, j.k. ( ). a multilevel memetic approach for improving graph k-partitions. ieee transactions on evolutionary computation, ( ), – , . benlic, u., & hao, j.k. ( a). breakout local search for maximum clique problems. computers & operations research, ( ), – . benlic, u., & hao, j.k. ( b). breakout local search for the max-cut problem. engineering applications of artificial intelligence, ( ), – . benlic, u., & hao, j.k. ( c). breakout local search for the quadratic assign- ment problem. applied mathematics and computation ( ), - . benlic, u., & hao, j.k. ( d). breakout local search for the vertex separator problem. in f. rossi (ed.) proceedings of the th intl. joint conference on artificial intelligence (ijcai- ), ijcai/aaai press, pages – , bei- jing, china, august . blum, c., puchinger, j., raidl, g.r., & roli, a. ( ). hybrid metaheuristics in combinatorial optimization: a survey. applied soft computing ( ), – . boese, k.d. ( ). cost versus distance in the traveling salesman problem. technical report tr- , ucla cs department. bontouxa, b., artigues, c., & feillet, d. ( ). a memetic algorithm with a large neighborhood crossover operator for the generalized traveling salesman problem. computers & operations research, , – . burkard, r.e. ( ). locations with spatial interactions: the quadratic assign- ment problem. in p. mirchandani and l. francis (eds.), discrete location theory, new york. cheng, c.h. ho, s.c., & kwan, c.l. ( ). the use of meta-heuristics for airport gate assignment. expert systems with applications, ( ), – . dorne, r., & hao, j.k. ( ). a new genetic local search algorithm for graph coloring. in a. e. eiben, t. bck, m. schoenauer, h.p. schwefel (eds.): ppsn , lecture notes in computer science, – . drezner, z. ( ). a new genetic algorithm for the quadratic assignment prob- lem. informs journal on computing, ( ), – . drezner, z. ( ). extensive experiments with hybrid genetic algorithms for the solution of the quadratic assignment problem. computers & operations research, ( ), – . duman, e., & or, i. ( ). the quadratic assignment problem in the con- text of the printed circuit board assembly process. computers & operations research, ( ), – . eshelman, l.j. ( ). the chc adaptive search algorithm: how to have safe search when engaging in nontraditional genetic recombination. proceedings of the first workshop on foundations of genetic algorithms, pages – , morgan kaufmann. fleurent, c., & ferland, j.a. ( ). genetic hybrids for the quadratic assign- ment problem. dimacs series in mathematics and theoretical computer science, – . garey. m., & johnson, d. ( ). computers and intractability. freeman. n.y.. hao, j.k. ( ). memetic algorithms in discrete optimization. in f. neri, c. cotta and p. moscato (eds.) handbook of memetic algorithms. studies in computational intelligence , chapter , pages – . hart, w.e., krasnogor, n., & smith, j.e. ( ). recent advances in memetic algorithms. studies in fuzziness and soft computing , springer. hassin, m., levin, a., & sviridenko, m. ( ). approximating the minimum quadratic assignment problems. acm transactions on algorithms, ( ), ar- ticle number . james, t. rego, c., & glover, f. ( a). a cooperative parallel tabu search algorithm for the quadratic assignment problem. european journal of oper- ational research, ( ), – . james, t. rego, c., & glover, f. ( b). multistart tabu search and diversi- fication strategies for the quadratic assignment problem. ieee transactions on systems, man, and cybernetics - part a: systems and humans, ( ), – . jones, t., & s. forrest, s. ( ). fitness distance correlation as a measure of problem difficulty for genetic algorithms. proceedings of the th international conference on genetic algorithms, morgan kaufmann, – . kelly, j.p., laguna, m., & glover, f. ( ). a study of diversification strategies for the quadratic assignment problem. computers & operations research, ( ), – . krasnogor, n., & smith, j. ( ). a tutorial for competent memetic algo- rithms: model, taxonomy and design issues. ieee transactions on evolu- tionary computing, ( ), – . li, f., xu, l.d., jin, c., & wang, h. ( ). random assignment method based on genetic algorithms and its application in resource allocation. expert systems with applications, ( ), – . lü, z., & hao, j.k. ( ). a memetic algorithm for graph coloring. european journal of operational research, ( ), – . merz, p., & freisleben, b. ( ). fitness landscape analysis and memetic algorithms for the quadratic assignment problem. ieee transactions on evolutionary computation, ( ), – . miao, z., cai, s., & xu. d. ( ). applying an adaptive tabu search algorithm to optimize truck-dock assignment in the crossdock management system. ex- pert systems with applications, ( ), – . misevicius, a. ( ). an improved hybrid genetic algorithm: new results for the quadratic assignment problem. knowledge based systems, ( - ), – . misevicius, a., & kilda, b. ( ). iterated tabu search: an improvement to standard tabu search. information technology and control, ( ), – . moscato, p. ( ). on evolution, search, optimization, genetic algorithms and martial arts: towards memetic algorithms. caltech concurrent computation program, report , california institute of technology, pasadena, califor- nia, usa. moscato, p., & cotta, c. ( ). a gentle introduction to memetic algorithms. in f. glover and g. a. kochenberger (eds.), handbook of metaheuristic. kluwer, norwell, massachusetts, usa. neri, f., cotta, c., & moscato, p. (eds.) ( ). handbook of memetic algo- rithms. studies in computational intelligence , springer. nikolić, m., & teodorović, d. ( ). a simultaneous transit network design and frequency setting: computing with bees. expert systems with applications, ( ), – . pardalos, p.m., rendl, f., & wolkowicz, h. ( ). the quadratic assignment problem: a survey and recent developments. in proceedings of the dimacs workshop on quadratic assignment problems, volume of dimacs series in discrete mathematics and theoretical computer science, – . porumbel, d.c., hao, j.k., & kuntz, p. ( ). an evolutionary approach with diversity guarantee and well-informed grouping recombination for graph coloring. computers & operations research, ( ), – . reeves, c.r. ( ). landscapes, operators and heuristic search. annals of operations research, , – . skorin-kapov, j. ( ). tabu search applied to the quadratic assignment prob- lem. orsa journal on computing, , – . stützle, t. ( ). iterated local search for the quadratic assignment problem. european journal of operational research, ( ), – . taillard, e. ( ). robust taboo search for the quadratic assignment problem. parallel computing, , - . wilhelm, m.r., & ward, t.l. ( ). solving quadratic assignment problems by simulated annealing. iie transactions, ( ), – . appendix this appendix includes two tables (tables and ). table shows the best, average and worst result obtained by bma after independent runs for selected qaplib instances, within a time limit of hours ( hours for instance tho ) and minutes respectively. table shows the computational results of our bma algorithm on the set of easy qap instances of the qaplib. for all these instances, each run of bma can attain the best known result with a computing time ranging from second to at most . minutes. table : computational results of the proposed bma algorithm with a time limit of hours and minutes respectively on the set of hard instances from the qaplib. columns ‘% δ̄best’, ‘% δ̄avg’ and ‘% δ̄worst’ show respectively the percentage deviation of our best, average, and worst solution from the best-known solution (bkr) over all the trials; column ‘tavg(min)’ indicates the average computing time in minutes when the best solution in each trial was attained. problem time limit of hours time limit of minutes name bks % δ̄best % δ̄avg % δ̄worst tavg(m) % δ̄best % δ̄avg % δ̄worst tavg(m) tai a . ( ) . . . . ( ) . . . tai a . ( ) . . . . ( ) . . . tai a . ( ) . . . . ( ) . . . tai a . ( ) . . . . ( ) . . . tai a . ( ) . . . . ( ) . . . tai b . ( ) . . . . ( ) . . . tai b . ( ) . . . . ( ) . . . tai b . ( ) . . . . ( ) . . . tai b . ( ) . . . . ( ) . . . tai b . ( ) . . . . ( ) . . . sko . ( ) . . . . ( ) . . . sko . ( ) . . . . ( ) . . . sko . ( ) . . . . ( ) . . . sko a . ( ) . . . . ( ) . . . sko b . ( ) . . . . ( ) . . . sko c . ( ) . . . . ( ) . . . sko d . ( ) . . . . ( ) . . . sko e . ( ) . . . . ( ) . . . sko f . ( ) . . . . ( ) . . . wil . ( ) . . . . ( ) . . . tho . ( ) . . . . ( ) . . . table : computational results of our bma algorithm for the set of easy instances of the qaplib. column ‘bkr’ indicates the optimal/best-known result ever reported in the literature. column ‘#best′ indicates the number of times a best-known solution was found over executions; column ‘% δ̄avg’ shows the percentage deviation of our average solution from the best-known solution (bkr) over all the trials; column ‘tavg(sec)’ indicates the average computing time in seconds when the best solution in each trial was attained. instance bkr #best % δ̄avg tavg(sec) instance bkr #best % δ̄avg tavg(sec) bur a / . . tai a / . . bur b / . . tai a / . . bur c / . . tai a / . . bur d / . . tai a / . . bur e / . . rou / . . bur f / . . rou / . . bur g / . . lipa a / . . bur h / . . lipa b / . . tai c / . . lipa a / . . chr a / . . lipa b / . . chr b / . . lipa a / . . chr c / . . lipa b / . . chr a / . . lipa a / . . chr b / . . lipa b / . . chr c / . . lipa a / . . chr a / . . lipa b / . . chr b / . . lipa a / . . chr a / . . lipa b / . . chr b / . . lipa a / . . chr c / . . lipa b / . . chr a / . . lipa a / . . chr b / . . lipa b / . . chr a / . . tai a / . . els / . . tai a / . . esc a / . . tai a / . . esc b / . . tai a / . . esc c / . . nug / . . esc d / . . nug / . . esc e / . . nug / . . esc f / . . nug a / . . esc g / . . nug b / . . esc h / . . nug / . . esc i / . . nug / . . esc j / . . nug / . . esc a / . . nug / . . esc b / . . nug / . . esc c / . . nug / . . esc d / . . nug / . . esc e / . . nug / . . esc g / . . nug / . . esc h / . . nug / . . esc a / . . scr / . . esc / . . scr / . . had / . . scr / . . had / . . tho / . . had / . . tho / . . had / . . wil / . . had / . . tai b / . . kra a / . . tai b / . . kra b / . . tai b / . . kra / . . tai b / . . ste a / . . tai b / . . ste b / . . tai b / . . ste c / . . sko / . . rou / . . sko / . . tai b / . . sko / . . tai b / . . sko / . . sign in skip to main content sign in » warwick.ac.uk you must sign in to view this page. your version of my warwick for ios is outdated. due to a bug affecting this version, you may be unable to login. please download the latest update from the app store. username password keep me signed in until i close my browser all day for three months indefinitely until i sign out   sign in find my account acceptable use policy address https://warwick.ac.uk/fac/sci/dcs/research/pcav/publications/pubs/eswa- .pdf service warwick.ac.uk © mmxxi terms privacy cookies accessibility visual privacy protection methods: a survey josé ramón padilla-lópeza,∗, alexandros andre chaaraouia, francisco flórez-revueltab adepartment of computer technology, university of alicante, p.o. box , e- alicante, spain bfaculty of science, engineering and computing, kingston university, penrhyn road, kt ee, kingston upon thames, united kingdom abstract recent advances in computer vision technologies have made possible the de- velopment of intelligent monitoring systems for video surveillance and ambient- assisted living. by using this technology, these systems are able to automatically interpret visual data from the environment and perform tasks that would have been unthinkable years ago. these achievements represent a radical improve- ment but they also suppose a new threat to individual’s privacy. the new capabilities of such systems give them the ability to collect and index a huge amount of private information about each individual. next-generation systems have to solve this issue in order to obtain the users’ acceptance. therefore, there is a need for mechanisms or tools to protect and preserve people’s privacy. this paper seeks to clarify how privacy can be protected in imagery data, so as a main contribution a comprehensive classification of the protection methods for visual privacy as well as an up-to-date review of them are provided. a sur- vey of the existing privacy-aware intelligent monitoring systems and a valuable discussion of important aspects of visual privacy are also provided. keywords: visual privacy protection, video surveillance, ambient-assisted living, computer vision, image processing . introduction it can be observed that world population is ageing. in fact, it is estimated that population over will rise by % between and , and those over will triple by (ec, ). furthermore, the number of people in long-term care living alone is expected to increase by % in japan, % in france and % in the us (ec, ). therefore, this situation will not be ∗corresponding author email addresses: jpadilla@dtic.ua.es (josé ramón padilla-lópez), alexandros@dtic.ua.es (alexandros andre chaaraoui), f.florez@kingston.ac.uk (francisco flórez-revuelta) preprint submitted to expert systems with applications february , sustainable in the near future, unless new solutions for the support of the older people, which take into account their needs, are developed. ambient-assisted living (aal) aims to provide a solution to this situation. aal applications use information and communication technologies to provide support to people so as to increase their autonomy and well-being. video cam- eras are being used more and more frequently in aal applications because they allow to get rich visual information from the environment. the advances pro- duced in the last decades have contributed to this. the computational power has been increased while, at the same time, costs have been reduced. fur- thermore, computer vision advances have given video cameras the ability of ‘seeing’, becoming smart cameras (fleck & strasser, ). this has enabled the development of vision-based intelligent monitoring systems that are able to automatically extract useful information from visual data to analyse actions, activities and behaviours (chaaraoui et al., b), both for individuals and crowds, monitoring, recording and indexing video bitstreams (tian et al., ). by installing networks of cameras in people homes or care homes, novel vision- based telecare services are being developed in order to support the older and disabled people (cardinaux et al., ). but these new technologies also sup- pose a new threat to individual’s privacy. traditionally, cameras are used in public spaces for surveillance services in streets, parking lots, banks, airports, train stations, shopping centres, museums, sports installations and many others. it is estimated that there is an average of one camera for every citizens in the uk, one of the most camera-covered countries in the world (gerrard & thompson, ). in short, video cameras are mainly used in outdoor environments and in public places, but they are not commonly used within private environments due to people’s concerns about privacy. generally, the use of video cameras in public places has been tolerated or accepted by citizens, whereas their use in private spaces has been refused. there may be several reasons to explain this difference. on the one hand, the perceived public-safety benefits favor the usage of cameras in public places for crime prevention, fight against terrorism and others. on the other hand, there is a widespread belief that while staying in public environments, people’s sensitive information will not be exposed. finally, there are some attitudes which have also contributed to accept their use, for example, to assume that anyone demanding privacy must have something to hide (caloyannides, ). in traditional video surveillance systems cameras are managed by human operators that constantly monitor the screens searching for specific activities or incidents. as estimated by dee & velastin ( ), the ratio between human operators and screens is around displays for every operator in four local authority installations within the uk. although they can only really watch - screens at once (wallace & diffley, ), this does not prevent abuses of these systems by their operators. furthermore, the processing capacities of next-generation video surveillance systems and the increasing number of closed- circuit television cameras installed in public places are raising concerns about individual’s privacy in public spaces too. in the near future it is expected that cameras will surround us in both public and private spaces. intelligent monitoring systems threaten individual’s right to privacy because of automatic monitoring (adams & ferryman, ). these systems can retain a variety of information about people habits, visited places, relationships, and so on (coudert, ). it is known that some systems already use facial recognition technology (goessl, ). this way, these systems may build a profile for each citizen in which the people identity and related sensi- tive information is revealed. therefore, this evolution of intelligent monitoring systems could be seen as approaching an orwellian big brother, as people may have the feeling of being constantly monitored. in the light of the above, it is clear that the protection of the individual’s privacy is of special interest in telecare applications as well as in video surveil- lance, regardless whether they operate in private or public spaces. therefore, privacy requirements must be considered in intelligent monitoring systems by design (langheinrich, ; schaar, ). as aforementioned, smart cameras become essential for aal applications. given that security and privacy protec- tion have become critical issues for the acceptance of video cameras, a privacy- aware smart camera would make it possible to use video cameras in realms where they have never been used before. if individual’s privacy can be guaran- teed through the use of this technology, public acceptance would be increased giving the opportunity of installing these cameras in private environments to replace simpler binary sensors or, most importantly, to develop new telecare services (olivieri et al., ; chen et al., ; morris et al., ). this breakthrough could open the door to novel privacy-aware applications for am- bient intelligence (ami) (augusto et al., ), and more specifically in aal systems for ageing in place (o’brien & mac ruairi, ), being beneficial to improve the quality of life and to maintain the independence of people in need of long-term care. . . related studies the focus of this review is on the protection of visual privacy. there are valuable reviews about ami and aal that have already considered privacy in video and have also highlighted its importance for the adoption of video-based aal applications (cook et al., ; cardinaux et al., ). but these works scarcely go into detail about how visual privacy protection can be achieved. other works from the video surveillance field have also analysed this topic but from a different point of view (senior et al., , ; cavoukian, ). the main threats and risks of surveillance technologies like closed-circuit television cameras, number plates recognition, geolocation and drones are discussed in depth. as a consequence, some guidelines to manage privacy are also proposed, but how to protect visual privacy is not considered. in the same line, senior & pankanti ( ) unify their previous works and extend the review of visual privacy not only to video surveillance but also to medical images, media spaces and institutional databases. they consider some technologies to protect visual privacy and provide a classification. as far as we know, this is the first attempt to provide such a classification of protection methods for visual privacy but it is not a comprehensive one. in a more recent work (winkler & rinner, ), security and privacy in visual sensor networks are reviewed. although they per- form a detailed analysis of the security from several points of view (data-centric, node-centric, network-centric and user-centric), they do not provide an in-depth analysis of privacy protection. in this survey, we focus on giving an answer to the question of how the vi- sual privacy can be protected, and how such a kind of protection is developed by some of the existing privacy-aware intelligent monitoring systems that have been found in the literature. because of this, a comprehensive classification of visual privacy protection methods is provided as the main contribution of this work. the remainder of this paper is organised as follows: section gives an intuitive notion of what visual privacy protection is. a comprehensive review of visual privacy protection methods is presented in section . in section , relevant privacy-aware intelligent monitoring systems are introduced. a discus- sion of important privacy-related aspects is carried out in section . finally, a summary of the present work as well as future research directions are given in section . . visual privacy protection privacy protection consists in preventing that the information that an in- dividual wants to keep private becomes available to the public domain. in the context of images and videos, we refer to it as visual privacy protection. in this paper, the terms visual privacy and privacy will be used indistinctly, except when indicated. first of all, it is worth to clarify when individual’s privacy needs to be protected. when protecting privacy, it can be differentiated between person’s identity and sensitive information which has to be kept in private. video can convey an enormous amount of information that can be qualified as sensitive. nevertheless, if sensitive information is present in a video but person’s identity is not, there is no privacy loss. the same is true whether person’s identity is in a video but without any sensitive information. in both cases, privacy is protected because there does not exist any association or mapping between sensitive information and person’s identity. another important issue related to visual privacy is which is the sensitive information or region of interest to be protected. in many works only the face is obscured but that is not enough to protect visual privacy. even when the person’s face is obscured, other elements could exist in the image through which person identification may be performed, for instance, using inference channels and previous knowledge (saini et al., ). visual cues like clothes, height, gait, and the like can be used to identify the person. for instance, in a pair- wise constraints identification (chang et al., ; chen et al., ) where faces had been masked, observers were able to identify whether a person in one image was the same one than in a different image. in that test, recognition hit rate was higher than %. by using this information and only detecting an image where a privacy breach exists, the person may be identified and tracked in images where privacy was presumably preserved. these visual clues must be considered in order to protect privacy as they affect to the election of which regions of interest have to be protected. so, there is not actually only one region of interest but multiple. a region of interest should be extended to a wider area in some cases, while two or more regions of interest should be created in others. . protection methods there are different ways to protect and preserve the privacy of individuals appearing in videos and images (see table ). two approaches can be found if we consider the temporal aspect of when a protection method is used, i.e. before the image is acquired, or after it. on the one hand, it is possible to prevent others of successfully capturing images in which an individual appears. on the other hand, once an image exists sensitive or private information (e.g. faces, number plates, and so on) can be removed (figure ). according to how privacy is protected, protection methods can be classified in five large categories: intervention (section . ), blind vision (section . ), secure processing (section . ), redaction (section . ) and data hiding (sec- tion . ). although the latter is often used along redaction methods, it has been added as it is a large field of research itself. in this section, a review of commonly used protection methods to protect visual privacy is presented. table : overview of reviewed papers according to the used visual privacy protection method. category count references intervention wagstaff ( ), patel et al. ( ), harvey ( ), mitskog & ralston ( ) blind vision avidan & butman ( ), erkin et al. ( ), avidan et al. ( ), sadeghi et al. ( ), shashanka ( ) secure processing erturk ( ), shashank et al. ( ), park & kautz ( ), ito & kiya ( ), upmanyu et al. ( ), ng et al. ( ), chaaraoui et al. ( a), zhang et al. ( ) redaction: image filter hudson & smith ( ), zhao & stasko ( ), boyle et al. ( ), neustaedter & greenberg ( ), kitahara et al. ( ), mart́ınez-ponte et al. ( ), neustaedter et al. ( ), zhang et al. ( ), frome et al. ( ), agrawal & narayanan ( ) continued on next page table – continued from previous page category count references redaction: encryption spanos & maples ( ), macq & quisquater ( ), agi & gong ( ), tang ( ), qiao et al. ( ), zeng & lei ( ), yang et al. ( ), boult ( ), yabuta et al. ( ), yabuta et al. ( ), dufaux & ebrahimi ( ), dufaux et al. ( ), chattopadhyay & boult ( ), baaziz et al. ( ), raju et al. ( ), dufaux & ebrahimi ( a), dufaux & ebrahimi ( b), xiangdong et al. ( ), carrillo et al. ( ), dufaux & ebrahimi ( ), tong et al. ( ), tong et al. ( ), dufaux ( ), sohn et al. ( ), pande & zambreno ( ), li et al. ( ), ra et al. ( ) redaction: k-same family newton et al. ( ), gross et al. ( a), gross et al. ( b), bitouk et al. ( ), gross et al. ( ), de la hunty et al. ( ), lin et al. ( ) redaction: object / people removal kokaram et al. ( ), igehy & pereira ( ), masnou & morel ( ), morse & schwartzwald ( ), efros & leung ( ), bertalmio et al. ( ), efros & freeman ( ), criminisi et al. ( ), criminisi et al. ( ), zhang et al. ( b), chan et al. ( ), cheung et al. ( ), shiratori et al. ( ), wexler et al. ( ), patwardhan et al. ( ), whyte et al. ( ), vijay venkatesh et al. ( ), koochari & soryani ( ), he et al. ( ), ma & ma ( ), ghanbari & soryani ( ), abraham et al. ( ) redaction: visual abstraction hodgins et al. ( ), lyons et al. ( ), tansuriyavong & hanaki ( ), fan et al. ( ), williams et al. ( ), chen et al. ( ), baran & popović ( ), hogue et al. ( ), chinomi et al. ( ), chen et al. ( ), qureshi ( ), sadimon et al. ( ), borosán et al. ( ), chen et al. ( ) data hiding petitcolas et al. ( ), wu ( ), cox et al. ( ), yabuta et al. ( ), zhang et al. ( a), ni et al. ( ), yu & babaguchi ( ), paruchuri & cheung ( ), cheung et al. ( ), cheung et al. ( ), paruchuri et al. ( ), guangzhen et al. ( ) . . intervention intervention methods deal with the problem of preventing someone to cap- ture private visual data from the environment. they aim to create capture- resistant spaces. these methods physically intervene camera devices to prevent the acquisition of an image by means of a specialised device that interferes with the camera optical lens. for instance, patel et al. ( ) developed the blindspot system. it locates any number of retro-reflective ccd or cmos cam- era lenses around a protected area and, then, it directs a pulsing light at the detected lens, spoiling any images that cameras may record as figure shows. figure : some privacy protected sequences of images: the first column shows the sensitive information or region of interest; the second column shows the region of interest replaced by the silhouette; the third column shows the sensitive information scrambled; and the last column shows the sensitive areas inpainted. reprinted from paruchuri et al. ( ). similarly, harvey ( ) proposed an anti-paparazzi device. it uses an array of high-power leds to produce a stream of light at over lumen. mitskog & ralston ( ) have patented a camera blocker for a device with an integrated camera that uses a thin film organic polymer to neutralise camera lens. this camera blocker is reusable and adhesively sticks to any surface without leaving figure : light beam neutralizing a camera lens. a light beam of single colour is used on the left picture, whereas a light beam generated using colour patterns is used on the right picture. reprinted from patel et al. ( ). a sticky residue. aforementioned approaches are suitable when, once recorded, no control can be enforced over the use of images. however, when enforcement is possible, software-based methods can be used. for instance, the firmware of a camera or an application installed on a smartphone could be responsible of preventing the capture of certain environments, such as artworks in a museum, when the device comes into the range of a bluetooth transmitter (wagstaff, ). nev- ertheless, software-based intervention has some drawbacks. first of all, it can be easily thwarted by using a camera without any privacy software installed. furthermore, users consent and their collaboration is required in order to work successfully. because systems like these depend on third parties, they are likely doomed to failure. a new privacy legislation that enforces cameras to be in accordance with privacy protocols is needed. anyway, given that no system is completely secure by nature, even under this assumption, it might be hacked. . . blind vision blind vision has to do with image or video processing in an anonymous way. it addresses privacy related processing issues by means of secure multi-party computation (smc) techniques that are applied to vision algorithms (avidan & butman, ). smc is a subfield of cryptography that enables multiple parties to jointly compute a function in such a way that their inputs and the function itself are not revealed. when applied to vision this means that one party, bob, could use a vision algorithm of another one, alice, without enabling bob to learn anything about alice’s algorithm and, in the same extent, without enabling alice to learn anything about bob’s images. these methods are useful in order to use third-party algorithms to process data in a privacy-respectful way. for instance, using blind vision techniques, common tasks such as face detection, image matching, object tracking or image segmentation could be done in an anonymous way. concerning this, avidan & butman ( ) developed a secure classifier that was used in a blind face detector. this classifier is based on an oblivious transfer method and a secure dot-product protocol in order to compute the face detection algorithm. however, it is very time consuming due to the high computational load of the algorithm. nevertheless, a proposal to accelerate the algorithm by using histograms of oriented gradients is considered at the risk of revealing some information about the original image. similarly, erkin et al. ( ) proposed an efficient algorithm that allows to jointly run the eigenfaces (turk & pentland, ) recognition algorithm. this algorithm operates over homomorphically encrypted data. it allows matching an encrypted facial image with a database of facial templates in such a way that the biometric itself and the detection result are kept in secret from the server that performs the matching. sadeghi et al. ( ) improved the communication and computation efficiency of the previous algorithm. blind vision algorithms are also used for image matching. avidan et al. ( ) presented a protocol for image matching related algorithms. image matching is described as a generalisation of detection and recognition tasks. the described algorithm is built upon a secure fuzzy matching of sift at- tributes, which are treated as character text strings, and it is used in their bag-of-features approach. a framework for privacy-preserving gaussian mixture model (gmm) com- putations is proposed by shashanka ( ). gmms are commonly used in machine learning for clustering and classification. in computer vision they are mainly used in background subtraction (zivkovic, ) for motion anal- ysis (chan & liu, ). . . secure processing there are other methods that are not based on smc but that process visual information in a privacy respectful way. they are referred as secure processing in this work. for example, an image matching algorithm for private content- based image retrieval (pcbir) is proposed by shashank et al. ( ). pcbir is related with similarity search. conversely to blind vision, privacy is required in one direction because the whole database is usually public, while the query is kept private. another possibility is to work with images in a domain that does not reveal visual information. ito & kiya ( ) presented an image matching algorithm using phase-only correlation (poc) in the frequency domain. this algorithm preserves the visual privacy of the images in a template database. in order to achieve this, all the images of the template database are converted to the frequency domain. then, a phase scrambling using an one-time key is applied to the discrete fourier transformation (dft) coefficients. afterwards, in order to match a query image with an image of the templates database, the query image is converted to the frequency domain and the matching is done with poc using the dft coefficients as inputs. algorithms that preserve privacy in an implicit manner, for instance, re- jecting visual information that is not necessary for the algorithm to work are also considered under the umbrella of secure processing. ng et al. ( ) pro- posed a privacy-preserving stereoscopic vision algorithm. it preserves privacy by calculating the disparity map using one-bit transform (erturk, ). this way, each pixel in the input images is reduced to one bit. in this process, a huge amount of information (colour, texture and so on) is removed in the out- put images, complicating thus the identification of persons appearing on the images. a prototype of a privacy-preserving system for recognition of activities of daily living is proposed by park & kautz ( ). this prototype relies on the silhouette of detected foreground objects and the motion-map of the frame in order to analyse the activities. this way, if the silhouette and motion-map gen- eration is performed within the camera and it is not possible to access the rgb signal, the visual privacy of the persons inhabiting the environment would be ensured from the algorithm point of view. similarly, chaaraoui et al. ( a) use silhouettes in their efficient approach for multi-view human action recogni- tion based on bag of key poses. by using silhouettes, identifiable information is removed, hence it can also be considered a privacy-aware method. depth infor- mation from rgb-d cameras can be used as a way to preserve privacy (zhang et al., ) as well. depth data can be obtained from low-cost structured-light cameras like microsoft kinect or asus xtion, or time-of-flight cameras like mi- crosoft kinect . given that no colour information is involved, a depth map visualisation does not enable face recognition and prevents direct extraction of visual clues for person identification. upmanyu et al. ( ) presented an interesting approach where an efficient framework to carry out privacy-preserving surveillance is proposed. this so- lution is inspired in a secret sharing scheme adapted to image data. in this framework an image is split into a set of random images in such a way that the shares do not contain any meaningful information about the original image. de- spite of shares being distributed between a certain number of servers, computer vision algorithms can be applied securely and efficiently. . . redaction image and video modification or redaction methods are the most common vi- sual privacy protection methods. they modify the sensitive regions of an image such as faces, bodies, number plates, etc. to conceal private information con- cerning the subjects appearing on it. in order to determine the privacy sensitive regions in which a redaction method operates, computer vision algorithms are used. however, this section focuses only on the application of privacy-preserving methods, therefore it is assumed that sensitive regions are correctly detected. according to the way in which an image is modified, redaction methods can be classified in several categories. to begin with, there are ad-hoc distor- tion/suppression methods (section . . ). these methods modify the region of interest of an image, either completely removing sensitive information from the image, or modifying the information using common image filters like blurring or pixelating. by using these filters, sensitive regions are modified in order to make them unrecognisable. more robust methods like image encryption (section . . ) are also used to conceal the region of interest. using image encryption techniques the privacy sensitive region of an image is ciphered by a key. generally encryption tech- niques based on scrambling (permutation) were commonly used in analogue video, but they can also be used in digital video. besides of that, it must be taken into account that in the literature scrambling is used as a synonym of encryption in some cases. in this paper, we consider scrambling techniques as a subfield of image encryption. another approach to image redaction is face de-identification (section . . ). these methods are focused on de-identifying the faces appearing in an image. although the aforementioned methods can be used for face de-identification, we focus here on those based on the k-same family of algorithms that implements the k-anonymity protection model (sweeney, ). these algorithms alter the face of a person in such a way that identity cannot be recognised but facial expressions are preserved. finally, some approaches like object removal (section . . ) use inpainting- based algorithms to completely remove sensitive regions of an image by filling the left gap with the corresponding background. once the image has been inpainted, a visual abstraction of the removed sensitive information can be rendered, like a stick figure, a point, a silhouette, and the like (chinomi et al., ). regarding image modification, some questions have to be outlined. redac- tion methods cannot modify an image in whatever way. as reported by hudson & smith ( ), there is a trade-off between providing privacy and intelligibil- ity in images. for instance, when an image is modified, information needed for image understanding may be also removed. so, a modified image could lack of utility and balancing privacy and intelligibility is needed. in other words, privacy protected images have to retain useful information needed by applica- tions built upon this information, such as telecare monitoring systems or video surveillance. it is also worth mentioning that whereas some of the redaction methods are irreversible, i.e. the modification cannot be undone, there are others methods that are reversible so the original version can be recovered if needed. nevertheless, reversible methods can be developed by using irreversible redaction methods along with data hiding algorithms. an overview of redaction methods has been given so far. in next subsections, each one of the described categories will be dealt with in depth, focusing on the most representative works and summarising how the presented methods are used. . . . image filtering redaction methods based on image filtering use common image filters to apply various effects on images in order to modify privacy sensitive regions (figure ). depending on the application, image filters can be used for ob- scuring human faces, human bodies, number plates or even background in video conferences (kitahara et al., ; mart́ınez-ponte et al., ; zhang (a) lena (b) lena with gaussian blur (c) lena pixelation figure : several examples of image filtering methods: a) the original image without any modification, b) image modified by applying a gaussian blur filter, and c) image modified by applying a pixelating filter. et al., ; frome et al., ). among the most commonly used filters, blur- ring (neustaedter & greenberg, ; zhang et al., ; neustaedter et al., ; frome et al., ) and pixelating (boyle et al., ; kitahara et al., ; neustaedter et al., ) stand out. a blurring filter applies a gaussian function over an image. this function modifies each pixel of an image using neighbouring pixels. as a result, a blurred image is obtained in which the details of sensitive regions have been removed. blurring is widely used in google street view (frome et al., ) to modify human faces and number plates. a pixelating filter divides an image into a grid of eight-pixel wide by eight-pixel high blocks. the average colour of the pixels of each block is computed and the resultant colour is assigned to all of the pixels belonging to that block. as a result, an image where the resolution of sensitive regions have been reduced is obtained. pixelating is commonly used in television to preserve the anonymity of suspects, witnesses or bystanders. however, it is vulnerable to some kind of attacks such as integrating pixels along trajectories over time that may allow partly recovering of the obscured information. lander et al. ( ) evaluated the effectiveness of pixelating and blurring in videos and static images. results indicated that participants were still able to recognise some of the faces. similarly, newton et al. ( ) showed that pixelating and blurring alike filters do not thwart recognition software. training a parrot recogniser using the same distortion as the probe on gallery images, high recognition rates are obtained (near %) despite looking somewhat de- identified to humans. . . . encryption of videos and images image and video encryption methods encode imagery data in such a way that the original data becomes unintelligible as can be observed in figure . the main goal of these methods is reliable security in storage and secure transmis- sion of content over the network. usually, security in cryptographic algorithms resides in the strength of the used key instead of in keeping the algorithm pri- figure : two examples of an encrypted image where the face of the person is considered the sensitive region. reprinted from boult ( ). vate. by using encryption, a distorted video that unauthorised viewers cannot visualise is obtained. only users who have the proper key for decryption can visualise it. through the inverse operation and the used key, the ciphered data can be deciphered in order to retrieve the original images. this is named condi- tional access through encryption. encryption methods operate over the whole frame or a delimited region of all the video frames. although such methods do not provide a balance between privacy and intelligibility, they enable to perform data analysis over unprotected data once authorisation and required permissions have been granted. in such cases, privacy would be protected until access to raw data is eventually requested. generally näıve video encryption algorithms have treated the compressed video bitstream as text data, therefore encrypting the entire video bitstream. hence, commonly used encryption algorithms such as data encryption standard (des), rivest’s cipher (rc ), advanced encryption standard (aes), rivest, shamir and adleman (rsa) and so on, have been used. these algorithms guarantee the highest security level but, unfortunately, they are not suitable for real-time video encryption because they are very time consuming (yang et al., ; pande & zambreno, ). due to this, selective encryption algorithms have been proposed (spanos & maples, ). these algorithms keep using text- based encryption but encrypt only a selected part of the video bitstream so as to get real-time encryption. other encryption algorithms have also been proposed for real-time encryption, namely light-weight encryption algorithms (zeng & lei, ). these algorithms are suitable for real-time applications because, when encrypting, they use a simple xor cipher or only encrypt some bits of the video bitstream. thereby, they are much faster than the first ones. finally, there are methods based on scrambling (tang, ). traditional scrambling methods modify an analogue video signal like those found on closed-circuit television cameras to make it unintelligible. however, with the proliferation of digital video cameras, scrambling techniques are also applied to digital videos in the field of video encryption. mainly, scrambling algorithms are based on permutation only methods in which transformed coefficients are then permuted in order to distort the resulting image. each one of these methods can operate in a specific domain, like the spatial domain, frequency domain (transform domain) or code-stream domain (com- pressed video). furthermore, it is important to note that light-weight en- cryption and scrambling-based methods are less secure than näıve encryption. for instance, scrambling video in the spatial domain is subject to efficient at- tacks (macq & quisquater, ). generally these algorithms trade-off security for encryption speed. however, compared to blurring and pixelating, scram- bling approaches as in dufaux & ebrahimi ( a) are successful at hiding identity (dufaux & ebrahimi, ; dufaux, ). regarding when encryption is performed, there are several approaches: prior to encoding, after encoding or during encoding. each approach has advantages and disadvantages. prior-to-encoding encryption is a very simple method that works with the original image independently from the used encoding scheme. however, it significantly changes the statistics property of the video signal, resulting in a less efficient compression later. regarding after-encoding encryp- tion, the compressed code-stream is encrypted after video encoding. the result- ing encrypted and compressed code-stream could hardly be reproducible in a standard player, and it could even cause the player to crash. however, it avoids to fully decode and re-encode the video. finally, during-encoding encryption has the advantage of a fine-grained control over the encoding process but it is closely linked to the used video encoding scheme. next, some of the selective and light-weight encryption as well as scrambling methods found in the literature will be analysed. concerning selective encryption, aegis (spanos & maples, ) is an algo- rithm that uses des to encrypt only the i-frames of the mpeg video bit- stream. by ciphering i-frames, the needed b-frames and p-frames cannot be reconstructed either. however, the algorithm is not completely secure due to partial information leakage from the i-blocks in p and b frames (agi & gong, ). similarly, the video encryption algorithm proposed by qiao et al. ( ) works with i-frames. it divides them in two halves that are xored and stored in one half. one of the half is encrypted using des algorithm. although this algorithm is secure, it is not suitable for real-time applications. raju et al. ( ) analyse the distribution of the dct coefficients (dc and ac) of com- pressed mpeg videos in order to develop a computationally efficient and secure encryption scheme for real-time applications. dc and ac coefficients are man- aged differently regarding their visual influence, and electronic code block and cipher block chaining modes are interleaved to adapt the encryption process to the video data. the described scheme uses rc for encrypting dct coefficients. boult ( ) used des and aes to encrypt faces in jpeg images during com- pression in their privacy approach through invertible cryptographic obfuscation. the information required for the decrypting process is stored inside the jpeg file header. although this information can be publicly read, it cannot be used without the private key. chattopadhyay & boult ( ) used this technique for real-time encryption, using uclinux on the blackfin dsp architecture. simi- larly, an encryption scheme for jpeg images is used by ra et al. ( ). the jpeg image is divided into two parts, one public and one private. the first one is unaltered, whereas in the second one the most significant dc coefficients are encrypted during the encoding process after the quantisation step. this ap- proach is designed to obtain jpeg-compliant images that can be sent to photo sharing services under storage and bandwidth constraints. as for light-weight video encryption, zeng & lei ( ) presented an efficient algorithm for h that operates in the frequency domain. bit scrambling is used to transform coefficients and motion vectors during video encoding without affecting the compression efficiency. by using this method each frame of the re- sulting video is completely distorted. a cryptographic key is used to control the scrambling process, thereby authorised users will be able to undo the scrambling using the key. a similar video encryption algorithm for mpeg- is proposed by dufaux & ebrahimi ( ) where security is provided by pseudo-randomly inverting the sign of selected transform coefficients (frequency domain) corre- sponding to the regions of interest. the encryption process depends on a pri- vate key that is rsa encrypted and inserted in the stream as metadata. in this method the amount of distortion introduced can be adjusted from merely fuzzy to completely noisy. this method is deeply explained for the case of motion jpeg in (dufaux et al., ), and for the case of h /avc in (dufaux & ebrahimi, a). concerning the latter, tong et al. ( ) made a proposal to correct the drift error produced in h /avc during video encryption. the described method is also used by baaziz et al. ( ) in an automated video surveillance system. dufaux & ebrahimi ( b) presented an extension of their previous work based on code-stream domain scrambling. the scrambling is per- formed after the mpeg- encoding process directly on the resulting code-stream output by the camera. this way, it avoids to decode and re-encode the video, saving computational complexity. an encryption scheme for jpeg xr (srini- vasan et al., ) working in the frequency domain is proposed by sohn et al. ( ). it uses subband-adaptive scrambling for protecting face regions. con- cretely, different scrambling techniques are used for each subband: random level shift for dc subbands, random permutation for lp subbands, and random sign inversion for hp subbands. a different encryption scheme based on compressive sensing (cs) (candes et al., ; donoho, ) and chaos theory is proposed by tong et al. ( ). this method scrambles sensitive regions of a video by using block-based cs sampling on their transform coefficients during encoding. the scrambling process is controlled by a chaotic sequence used to form the cs measurement matrix. in order to prevent drift error they also use their previous method. finally, regarding scrambling, tang ( ) proposed an encryption algo- rithm that works in the frequency domain and uses the permutation of the dct coefficients in order to replace the zig-zag order. specifically, instead of using the zig-zag order to map the x block to a x vector, a random per- mutation list is used. a different approach is proposed by yabuta et al. ( ) where scrambling is performed in the space domain before encoding. the scram- bling only affects the moving regions and is performed by randomly permuting pixels. before being scrambled, the original moving regions are encrypted with aes, and embedded inside the masked image using digital watermarking (pe- titcolas et al., ). thereby, the scrambled regions can be recovered later if needed. an improved version of this method is presented in (yabuta et al., ) being able of real-time encoding, and decoding only one specific object by exploiting object tracking information. xiangdong et al. ( ) presented a novel encryption scheme that relies on chaos theory. it works in the space domain and prior to video encoding. it permutes the pixels of an image row by row by using a chaotic sequence of sorted real numbers as a key. this key can be used to de-scramble the image when necessary. a similar approach that sorts the chaotic sequence as vigenère cipher is proposed by li et al. ( ). carrillo et al. ( ) proposed an encryption algorithm working in the spatial domain. before encoding, pseudo-random permutations are applied to the pix- els. a secret pass phrase which controls the permutation process is used as key. this algorithm is independent of the used compression algorithm and robust to transcoding at the cost of an increase in bitrate depending on the percentage of encrypted blocks. . . . face de-identification face de-identification consists in the alteration of faces to conceal person identities. the goal is to alter a face region in such a way that it cannot be recognised using face recognition software. in order to achieve this, the faces appearing in images are modified by using some of the aforementioned methods such as image filtering previously discussed in section . . . nevertheless, there are cases in which privacy must be preserved while images must still keep their capacity of being analysed. in such cases, it is then necessary to balance privacy and intelligibility. concerning this, there are methods in the literature that consider this trade-off, and some of them will be reviewed in this section. the k-same family of algorithms (figure ) is one of the most recent and commonly used algorithms for face de-identification. k-same was first intro- duced by newton et al. ( ). intuitively, k-same performs de-identification by computing the average of k face images in a face set. then, all of the im- ages of the given cluster are replaced by the obtained average face. using this algorithm, a de-identified face is representative of k members of the original used face set. this way, if the probability of a de-identified face of being cor- rectly recognised by a face recognition software is no more than k , it is said that this algorithm provides k-anonymity privacy protection (sweeney, ). however, despite the fact that this formal model can preserve privacy, there are no guarantees of the utility of the data. gross et al. proposed some extensions to the k-same. on the one hand, an algorithm named k-same-select that extends the k-same algorithm is presented in (gross et al., a). it guarantees the utility of the data, for example, by preserving facial expressions or gender in face images. this algorithm di- figure : several examples of de-identified face images: a) faces de-identified by using the k-same algorithm where some ghosting artefacts appear due to misalignments in the face set. b) faces de-identified by using k-same-m algorithm. reprinted from gross et al. ( ). vides the face image set into mutually exclusive subsets using a data utility function and, then, it uses the k-same algorithm on the different subsets. on the other hand, the k-same-m algorithm is introduced in (gross et al., b) to fix a shortcoming of the k-same-select algorithm. appearance-based algo- rithms, like k-same-select, work directly in the pixel level producing sometimes alignment mismatch that leads to undesirable artefacts in the de-identified face images. in order to overcome this issue k-same-m relies on active appearance model (cootes et al., , ), a statistical and photo-realistic model of the shape and texture of faces. this model is also used by de la hunty et al. ( ) for real-time facial expression transferring from one’s person face to another. despite this method is not used for face de-identification, it could be useful to transfer expressions from an already de-identified face to another one. the k-same-based algorithms have some constraints. for instance, if a sub- ject is represented more than once in the dataset then k-same does not provide k-anonymity privacy protection. in order to address this shortcoming, gross et al. ( ) proposed a multi-factor model which unifies linear, bilinear and quadratic models. by using a generative multi-factor model, a face image is fac- torised into identity and non-identity factors. afterwards, the de-identification algorithm is applied and the de-identified face is reconstructed using the multi- factor model. a different approach is described by bitouk et al. ( ), where faces are automatically replaced in photographs. a large library of d faces downloaded from the internet is built according to appearance and pose. in order to replace a detected face in an image, a similar candidate to the input is selected from the library. lin et al. ( ) presented a similar work for face swapping where a personalised d head model from a frontal face is built, thereby any pose can be rendered by directly matching with the face that wants to be substituted. . . . object / people removal object and people removal deals with concealing persons or objects appear- ing in an image or a video in such a way that there are not trails of them in the resulting modified version (figure ). for the sake of clarity, the term ‘object’ will be used to refer both person and object indistinctly. when removing an object from an image a gap is left. this gap has to be filled in order to create a seamless image. inpainting methods are used to repair regions with errors or damages. inpainting consists in reconstructing missing parts in such a way that the modification is undetectable. information from surrounding area is used to fill in the missing areas. therefore, these methods can be used for visual privacy to remove people that do not commit suspicious activities from video surveillance recordings, removing inactive participants from the video stream of a video conference, concealing people from images in online social networks, and so many other applications. however, due to computational restrictions, inpainting methods are rarely used in real-time applications running on com- modity hardware (granados et al., ). inpainting can be divided into two large groups: image inpainting (bertalmio et al., ) and video inpainting (kokaram et al., ; abraham et al., ). this distinction is mainly due to the content nature. while in an image is only needed to ensure spatial consistencies, in video, temporal consistencies between all of the frames have to be ensured as well. regarding image inpainting, there are several approaches: texture synthe- sis (igehy & pereira, ; efros & leung, ; efros & freeman, ), partial differential equations (pde) inspired algorithms (masnou & morel, ; morse & schwartzwald, ; bertalmio et al., ; chan et al., ), and exemplar based (criminisi et al., , ; whyte et al., ; koochari & soryani, (a) real image (b) modified image figure : an example of a people removal method where the person has been manually selected in the real image (a), and then automatically removed in the second image (b) by filling the region concerning the person using an exemplar-based image inpainting method. reprinted from criminisi et al. ( ). ; he et al., ; ma & ma, ). in texture synthesis, a synthetic texture derived from one portion of the image is used to fix another portion. this syn- thetic texture is a plausible patch that does not have visible seams nor repetitive features. algorithms based on texture synthesis are able to fill in large regions but at the cost of not preserving linear structures. pde inspired algorithms re- construct the gap using geometry information to interpolate the missing parts. a diffusion process propagates linear structures of equal gray value (isophotes) of the surrounding area into the gap region. although these algorithms pre- serve well linear structures, the diffusion process introduces some blurring when filling in large regions. finally, exemplar-based methods are of particular in- terest. instead of generating synthetic textures, these methods operate under the assumption that the information that is necessary to complete the gap has to be fetched from nearby regions of the same image. they generate new tex- tures by searching for similar patches in the image with which the gap is filled. moreover, some exemplar-based algorithms also search the needed information in databases of millions of images in order to complete the remaining infor- mation (whyte et al., ). furthermore, these algorithms often combine the advances of texture synthesis and isophote-driven inpainting by a priority-based mechanism which determines the region filling order, thereby reducing blurring caused by prior techniques. (a) real image (b) blur (c) pixelating (d) emboss (e) solid silhouette (f) skeleton (g) d avatar (h) invisibility figure : several examples of visual abstractions where the person in the real image (a) has been replaced by a visual model. regarding video inpainting, some of the first straightforward approaches tried to apply image inpainting methods to individual images of the underlying video data. however, they did not take full advantage of the temporal correla- tion of video sequences. due to this, the previous methods are often modified in order to be adapted to sequences of images, reconstructing a given frame by interpolating missing parts from adjacent frames. these methods can be classified into patch-based methods (shiratori et al., ; wexler et al., ; patwardhan et al., ; ghanbari & soryani, ) and object-based meth- ods (zhang et al., b; cheung et al., ; vijay venkatesh et al., ). as patch-based methods are unable to perform both spatial and temporal aspects simultaneously, object-based methods were introduced in order to overcome these constraints. . . . visual abstraction / object replacement object replacement involves the substitution of objects (or persons) appear- ing in an image or video by a visual abstraction (or visual model) that protects the privacy of an individual while enabling activity awareness. as far as we know, the term ‘visual abstraction’ was early coined by chinomi et al. ( ) to refer to a visual model that abstracts a replaced object. common visual models could be a point, a bounding box, a stick figure, a silhouette, a polyg- onal model, and many others as seen in figure . visual abstraction may be obtained in a a variety of ways. hence, there is an overlap with some of the previous reviewed methods, such as image filtering or face de-identification, that could also be used as visual models. object replacement does not necessarily imply removing the object to be replaced, but quite often the aforementioned techniques in section . . intervene. in such cases, an object removal method is applied, followed by the rendering of the visual model over the inpainted image. the abstract object is often located in the same relative position, pose, and orientation as the original object. although it depends on the used visual model, by using a proper abstract representation of an object, the information of the scene remains useful so as visual analysis can still be carried out, for instance, to assess the underlying activity before an accident in a home risk detection service. then, the activity can be analysed without violating the right to privacy of people appearing in the image because it is not possible to directly identify them. however, as reported by hodgins et al. ( ), some works in human motion perception showed that viewers easily recognise friends by their gaits, as well as the gender of unfamiliar persons when using moving light displays (johansson, ). concerning this, it seems that motion is essential for identifying human figures. furthermore, apparently the geometric model used for rendering human motion affects the viewer’s perception of motion. for example, in experiments carried out by hod- gins et al., subjects were able to better discriminate motion variations using the polygonal model than they were with the stick figure model. hence, a study to determine how visual abstraction techniques actually preserve privacy must be done. different works employ visual abstraction and object replacement. for ex- ample, a silhouette representation can be used as a visual abstraction of a per- son (tansuriyavong & hanaki, ). this removes information about textures while maintaining the shape of the person, thereby complicating the identifi- cation. a silhouette representation is used, for instance, to preserve individ- ual’s privacy in a fall detector and object finder system (williams et al., ). another representation can be obtained by using an edge motion history im- age (chen et al., , ). by using this pseudo-geometric model, the whole human body is obscured and the person looks like a ghost in the final image. this method detects edges of the body appearance and motion, accumulating them through the time, in order to smooth the noise, and partially preserve body contours. chinomi et al. ( ) proposed twelve operators for visual abstrac- tions: as-is, see-through, monotone, blur, mosaic, edge, border, silhouette, box, bar, dot and transparency. in order to choose among one of them, the relation- ship between the subject being monitored and the viewer is taken into account. in addition to the representation that have been seen so far, d avatars can be used too. avatar creation (sadimon et al., ) is a very interesting field with straightforward application in visual abstraction as well. lyons et al. ( ) presented a method that uses automatic face recognition and gabor wavelet transform for creating avatars. it automatically extracts a face from an image, and create a personalised avatar by rendering the face into a generic avatar body. however, given that the avatar maintains recognisable aspects of the face, pri- vacy would not be protected. hogue et al. ( ) proposed a method based on d reconstruction for whole body avatar creation. they use a portable stereo video camera to extract the geometry of the person and create a d model. a similar and more recent work is proposed by chen et al. ( ), where two low- cost rgb-d cameras are used to scan the whole body of a person in order to create a mesh model. although the resulting textured model of both methods do not protect privacy, they can constitute a building block for other methods where generic d avatars are used to replace persons, as proposed by fan et al. ( ). furthermore, automatic rigging (baran & popović, ; borosán et al., ) could be used to animate d models, enabling the use of customised d models in real-time. hence, opening the door to new protection methods that transform or deform mesh models. finally, the use of object-video streams for preserving privacy is proposed by qureshi ( ). a separated video bitstream is generated for each foreground object appearing in the raw video. thereby, the original video can be reconstructed from object-video streams without any data loss. during reconstruction, each video bitstream can be rendered in several ways, for instance, obscuring people identities using a silhouette representation or just not showing an object-video stream at all. . . data hiding based methods in order to protect privacy, there are redaction methods that apart of mod- ifying the region of interest, they embed the original information inside of the modified version so as to be retrieved in the future if needed (cheung et al., ). these redaction methods make use of data hiding techniques (petitcolas et al., ) to develop reversible methods when the underlying redaction does not support it. in figure , a classification of data hiding methods is presented. concerning the terminology used in data hiding, the embedded data is the mes- sage that will be sent secretly. it is often hidden in another message referred to as cover message whose content can be text, audio, image or video. as a result of a hidden process, a marked message is obtained. generally data hiding techniques are used for steganography, digital water- marking and fingerprinting. steganography is a practice that consists in con- cealing a secret message as embedded data inside a cover message. the hiding process is often controlled by a key in order to allow the recovery of the secret message to parties that know it. regarding digital watermarking and finger- printing, they both use steganography techniques but are focused on copyright protection. on the one hand, digital watermarking encodes information about the owner of an object by means of a visible pattern (e.g. a company logo) or hidden information. on the other hand, fingerprinting is used to embed hidden serial numbers that uniquely identify an object in such a way that the owner of the copyright can detect violations of licence agreements. data hiding methods have to provide different kind of features in terms of capacity, perceptibility and robustness (cox et al., ). the main difference between steganography information hiding covert channels steganography linguistic steganography technical steganography anonymity copyright marking robust copyright marking fingerprinting watermarking imperceptible watermarking visible watermarking fragile watermarking figure : a classification of data hiding techniques according to petitcolas et al. ( ). and digital watermarking (and fingerprinting) is their application. whereas the former aims for imperceptibility to human vision, the latter is focused on the robustness. invisible watermarking can be used for privacy protection purposes. instead of embedding information about the owner, the original video is embedded in- side of the privacy protected version of it. due to this, data hiding methods that provide large embedding capacity are required. furthermore, if the reversibility of the hiding process is considered, irreversible and reversible data hiding tech- niques can be found (ni et al., ). in the former, the cover video cannot be fully restored after the hidden data is extracted, but it usually produces higher hiding capacity. in the latter, the cover video can be fully restored, thereby maintaining the authenticity of the original video. nevertheless, reversible data hiding is not required in privacy protection because the cover video is discarded once the embedded data has been extracted. some of the privacy protection methods that make use of non-reversible and imperceptible watermarking are described next. for instance, yabuta et al. ( ) extract moving objects of a motion jpeg video. these regions are scrambled in the original video, resulting in a privacy protected video that is used as the cover video. then, the extracted objects are jpeg compressed and encrypted with aes, followed by a hiding process that embeds them into the least significant bits of middle frequency dct coefficients of the cover video. the added perturbation is small and does not visually affect the reconstructed image. a similar approach is proposed by zhang et al. ( a), where fore- ground objects are extracted from a h video, compressed and encrypted as a regular video bitstream. the gaps left by the foreground objects are inpainted in the original video using background replacement. then, the encrypted fore- ground objects are embedded in the dct coefficients of the inpainted video considering the perceptual quality of the marked video. however, one drawback is the increased bitrate of the resultant video. yu & babaguchi ( ) presented a method for hiding a face into a new generated face. it considers the face as the information to be embedded. initially, an active appearance model is built and faces are characterised as parameters according to this model in order to reduce the payload of the embedded information. then, the face appearing in the orig- inal mpeg video is masked. finally, face parameters are embedded into dct coefficients of the video following the quantised index modulation scheme (cox et al., ). a novel data hiding algorithm for m-jpeg that minimises both the output perceptual distortion and the output bitrate is proposed by paruchuri & cheung ( ). the algorithm identifies the optimal locations to hide data by selecting the dct coefficients that minimise a cost function that considers both distortion and bitrate. the person appearing in a video (privacy data) is extracted and removed from it. then, the privacy data is embedded into the selected dct coefficients of the inpainted video. the proposed algorithm is used by cheung et al. ( ) but modifying the rate-distortion scheme to determine the optimal number of bits to be embedded in each dct block. this last version is used for a reversible data hiding scheme in a video surveillance prototype (cheung et al., ; paruchuri et al., ). guangzhen et al. ( ) presented a different method for jpeg images. it uses scaling and wavelet coef- ficients generated by dwt. the former is used to build a low-resolution image where privacy information appears pixelated, whereas the latter is used along scaling coefficients to jointly recover the original image. in this way, wavelet coefficients represent only the private information that needs to be embedded. they are embedded in the dct coefficients of the low-resolution image after quantisation via amplitude module modulation (wu, ). . privacy-aware intelligent monitoring systems this section introduces some of the existing expert and intelligent systems in the video surveillance and aal fields that take privacy into account and, thereby, are developed as a framework to protect it. some of the reviewed systems focus only on the way in which private date is managed, whereas others use redaction methods to protect individual’s privacy before imagery data is stored, distributed, shared or visualised. because of this, some privacy design questions which such systems should address are drawn (senior et al., ; yu et al., ): • what data is captured? • has the subject given consent? • how does the subject specify the privacy preferences? • what form does the data take? • who sees the data? • how long is the data kept? • how is the data used? • how raw is the data? • who and what should be protected? • how should privacy be protected in a video surveillance system? • how can privacy be protected without losing utility? the answers to these questions lead to privacy policies and technical aspects that such systems should cope with in one way or another. in the same line, it is also important to question when redaction is performed. as stated by senior ( ), there are mainly several locations where redaction can be performed for a general video-based monitoring architecture made up of cameras, video processor, database and user interfaces. concretely, these locations include all of the mentioned parts but cameras because of technical issues that may arise in legacy systems. nevertheless, cameras have been also included here from a conceptual point of view, so there are a total of four locations. in such an architecture, data flows from video cameras to user interfaces, crossing through the video processor and the database. furthermore, depending on privacy policies of subjects, viewers permission, and others, both redacted and unredacted data may have to be displayed. next, we enumerate the several locations where redaction can be applied: . user interface. this location is the most insecure. data crosses the system without being protected until it reaches the user interface. then, redac- tion is carried out by the user interface and the protected information is presented to the viewer. an advantage of this approach is that redacted data does not require to be stored. however, metadata information needs to be delivered with the raw data. . database. when the user interface requests the database for some infor- mation to view, the latter can redact it. thereby, redacted data does not require to be stored but it involves additional processing because the same data may be redacted multiple times. besides of that, latency issues may arise due to extra processing. . video processor. redaction can be performed by the video processor before storing. it analyses video bitstreams to detect activities and extract useful information. nevertheless, bandwidth and storage requirements are increased because both redacted and unredacted data need to be sent to the database. . video camera. a privacy-aware smart camera can redact sequences of images itself. this is the earliest possible stage in which redaction can be applied. similarly to the video processor, in this location bandwidth and storage requirements are increased too. several redaction locations can be combined to enhance security as in double redaction also proposed by senior ( ), in which privacy protection is applied at the earliest stage as well as in other locations. video is decomposed in mul- tiple information streams containing the private data which are encrypted and flow through the system to the database. the information can be recombined later, when needed by an authorised viewer. by using double redaction more than one level of privacy protection can be provided, where each one could be suitable for different applications. this way, some visual information could not be protected but securely stored, whereas other information could be redacted according to several factors like subject, viewer, ongoing activity and so on. next, some of the existing systems found in the literature are described. comedi (coutaz et al., ) is a media space prototype that facilitates remote informal communication and group awareness while assuring privacy protection. a porthole display with fish-eye feature is used in order to provide awareness of the remote activities that are being carried out. it shows the personal in- formation that the corresponding remote user has previously accepted to reveal without losing awareness about peripheral activities. regarding privacy, this prototype uses a face tracker and a privacy filter based on eigenspace coding in order to filter the captured faces not belonging to the image set of ‘socially correct’ faces. nest (fidaleo et al., ), the networked sensor tapestry, is a general ar- chitecture to manage the broad range of distributed surveillance applications. it is scalable and is composed of software modules and hardware components. the core component of a nest server is the privacy buffer. it utilises programmable plug-in privacy filters which operate on data coming from sensors in order to prevent access to it or to remove personal identifiable information. these pri- vacy filters are specified using a privacy grammar that is able of connecting multiple low-level privacy filters to create arbitrary data-dependent privacy def- initions. moreover, the privacy of the monitored subjects is also considered. in this sense, the nest architecture integrates individuals’ privacy (according to behaviours) denying access to specific information to some or all of the modules or operators. nest features secure sharing, capturing, distributed processing and archiving of surveillance data. stealth vision (kitahara et al., ) is an anonymous video capturing sys- tem that protects the privacy of objects by pixelating or blurring their appear- ance. it determines private areas of captured videos from a mobile camera using d space models. the d position of the target object is estimated by a ho- mographic transformation using images coming from a static overhead camera. afterwards, a calibrated mobile camera estimates the privacy area by project- ing the d models onto the captured image plane. finally, the privacy area is protected. rfid sensors are employed in order to indicate which persons or objects are allowed to be captured. prisurv (chinomi et al., ) is a system that uses visual abstract repre- sentations to protect individual’s privacy. it is composed of several modules: analyser, profile generator, profile base, access controller, abstractor and video database. in order to identify subjects, rfid tags and image processing is used. this system is focused on small communities where a certain number of members are registered and they monitor each other. given that the sense of privacy is different for everyone, privacy policies are determined according to the closeness between subjects in the video and viewers monitoring. hence, the appearance of subjects is opened to close viewers while it is hidden to the viewers that subjects feel distant from. depending on this closeness, different abstract representations are chosen. respectful cameras (schiff et al., ) is a prototype system focused on addressing privacy concerns. this system works by detecting coloured markers in real time that people wear such as hats or vests in order to indicate their privacy preferences. this are expressed as their will of remaining anonymous. the system obscures the face of a person when a marker has been detected. carelog and bufferware (hayes & truong, ) are two systems based on the concepts of the selective archiving model. under this model, data is con- stantly buffered but an explicit input is required in order to archive it, otherwise data is lost. it represents a compromise whereby people can negotiate their own policies concerning control, privacy, information access and comfort. carelog is a system aimed at education classrooms for the recording of diagnostic be- havioural data of children with severe behaviour disorders. using such a system, a teacher takes control of data archiving to document an incident after it has occurred. regarding bufferware, it is a system aimed at semi-public spaces in which people can save recorded data, if desired, using a touch-screen. they are also able of viewing previous recordings. in this case, bufferware was placed in a social area of an academic building. these two systems were proposed in order to conduct an experiment to assess issues of the selective archiving model, such as ownership of data, choice of which data should be saved or deleted, vis- ibility and awareness of recordings, and trust in the fact that policies are being followed and features were implemented as described. altcare (shoaib et al., ) is a monitoring system based on a static net- work of video cameras aimed for emergency detection at home. it focuses on fall detection, and works without any manual initialisation or interaction with older persons. when a fall is detected, altcare first attempts to get the confirmation from the involved person. afterwards, if the verification is positive, it auto- matically communicates the emergency to the responsible person. in addition to information concerning the involved person, the system transmits a patch of the video that shows the emergency. in order to protect the privacy, the person is replaced by the silhouette in the video. furthermore, altcare enables system administrators to check the state of the person at any time, if so desired by the monitored person. in sohn et al. ( ), a privacy-preserving watch list screening system for video-surveillance is proposed. depending on the group of interest the watch list comprise either blacklisted or white-listed identities. hence, the goal is to verify whether a person is enrolled in the watch list or not. in order to preserve privacy it relies on homomorphic encryption. this system is composed of a network of distributed cameras and a central management server (cms) which owns the watch list database. this database is considered private and it contains face feature vectors corresponding to identities of interest. in this system, cameras analyse the recorded raw video in order to detect face regions. when a face is detected, it is encrypted and sent to the cms where face feature vectors are retrieved from the watch list database. afterwards, a comparison between the face coming from a camera and those from the watch list is performed in the encryption domain. then a result confirming whether the identity is included in the watch list or not is encrypted and sent back to the camera. finally, according to the obtained result the camera notifies it to the security service. trustcam (winkler & rinner, b,a) is a smart camera that provides se- curity and privacy-protection based on trusted computing. by using this smart camera video bitstreams are digitally signed. thereby, the manipulation of recorded images or videos can be detected, origin of images are provided, and multi-level access control is supported. each trustcam node is operated by a central control station (cs) which runs a protected database where crypto- graphic keys generated during camera setup are securely stored. camera nodes encrypt privacy sensitive image data such as faces or number plates using a spe- cial cryptographic key stored in each one. moreover, an abstracted version of the sensitive regions as, for example, a silhouette can also be generated and en- crypted. these encrypted sensitive regions cannot be decrypted without having access to the cs. indeed, each representation requires a different key in order to decrypt it, thereby providing multi-level access control. finally, paas (barhm et al., ) is a privacy-aware surveillance system that enables monitored subjects to specify their privacy preferences via gestures. it uses face recognition technology in order to map individuals to their privacy preferences and security credentials that are encoded using an extension of the p p-appel framework (cranor et al., ). users have access to one of three privacy settings, namely no privacy, blurred face and blurred full body. . discussion . . recognition of the region of interest as it has been seen throughout this survey, privacy protection methods mainly focus on preserving visual privacy in images and videos, either modifying completely the whole image or only a region of interest. although the detection of the region of interest has not been specifically considered in this work, correct detection is fundamental in order to make protection methods work as desired. for instance, it could be imagined a hypothetical case in which a scrambling algorithm was used to obfuscate a person’s face in a video. supposing that the video sampling rate is fps, there are frames in minutes of video where a face appears. it could be supposed also that the hit rate of the used face detector was around %. under these assumptions, the subject face will be poorly detected in roughly frames of video, thereby not enabling the obscuring algorithm to work properly in protecting privacy. if the face is fully observable in a single frame, subject identification can already be performed. not only that, the face could also be inferred observing multiple frames in which the face detector failed. hence, it is essential that the computer vision techniques (segmentation, object detection, tracking, recognition, etc.) involved before the obscuring process are reliable and robust enough in order to guarantee the effectiveness of the protection method. . . privacy and intelligibility trade-off although we have mentioned the privacy-intelligibility trade-off in the intro- duction of the section . , let us provide a further discussion here. zhao & stasko ( ) compared several image filters (blurring, pixelating, edge-detector, live- shadow and shadow-view) and found out that whereas intelligibility is provided almost in all of the tested filters, actors were also recognised in most of the cases. boyle et al. ( ) performed an in depth evaluation of blurring and pixelating to analyse how the variation of the filtration level affects to this balance. the obtained results suggested that blurring and, to a lesser extent, pixelating may provide a balance, but it would be a precarious balance at best. furthermore, neustaedter et al. ( ) state that blurring by itself does not suffice for privacy protection when a balance is needed. therefore, it seems that image filters can hardly provide a balance between privacy and information utility, and generally privacy loses in this balance. nevertheless, when such a balance is not needed, image filters may be used. as reported by agrawal & narayanan ( ), blur- ring is enough to hide the identity if gait is not involved. but gait and other temporal characteristics are difficult to hide if there is some familiarity between the subject and the user. anyway, it would be interesting to expand these stud- ies to also include visual abstraction methods and all those where such a balance could be analysed. . . real-time applications regarding the techniques that could be used in real-time intelligent moni- toring systems, apart from image filtering, the remainder techniques would be impracticable because they are very time consuming. in such systems, all the involved computer vision algorithms along with the protection method must jointly work in real time. like inpainting techniques, encryption ones have high computational requirements. nevertheless, some light-weight encryption techniques may work properly. in turn, some image inpainting techniques are designed to work in real time too. the election of which protection method to use depends on the application requirements. . . privacy model and evaluation relying on a formal model to guarantee privacy preservation is required in order to have a common framework in which privacy solutions can be devel- oped and evaluated. common definitions about what visual privacy should be, when privacy is protected, which are the sensitive areas, and so on are needed. a model that currently fulfils some of these requirements is the face de- identification based on the k-same family of algorithms. although it guarantees that a de-identified face cannot be recognised with a probability higher than k , inference channels are not considered. in that sense, saini et al. ( ) propose a model that takes this into account. this model measures the privacy loss of the individuals that usually inhabitant a surveillance area. this measure is given as a continuous variable in the range [ , ]. as far as we know, this is the only approach that currently measures privacy loss in such a way. regarding the remainder methods, despite they are not based on a formal model, it can be intuitively realised that they provide different protection levels. for instance, näıve image filtering techniques like blurring or pixelating are not sufficient to protect privacy when intelligibility is involved. but blanking out or encrypting sensitive regions could be enough. also related to the previous point is how the evaluation of visual privacy pro- tection is performed. we have not found a common framework for it. although the aforementioned model for measuring privacy loss is very valuable, it cannot be used to build such a framework that automatically performs the comparison because current technology is not robust enough. this would have been very useful in order to perform an exhaustive comparison covering all of the reviewed methods. moreover, the lack of more manually-labelled privacy datasets also contributes to deteriorate visual privacy evaluation. the only datasets focused on privacy that have been found in the literature are pevid-hd (korshunov & ebrahimi, ) and pevid-uhd (korshunov & ebrahimi, ). both datasets consider video surveillance scenarios and provide high definition video sequences and ultra high definition ones, respectively. a similar dataset for ambient-assisted living would be appreciated where, in addition, other kind of visual sensors were included (e.g. rgb-d sensors). in addition to assess the grade in which privacy is protected, to what extent useful information is retained should also be studied. in the absence of such evaluation mechanisms, several works have been found in which some of the protection methods have been tested by users as part of the experimentation. these experiments are mainly based on conducting interviews and questionnaires with users where their skills in extracting useful information from a privacy-protected image (subject identification, pose, activity, presence, etc.) are evaluated (korshunov et al., ). results obtained from an objective study are missed. . . user studies about privacy requirements concerning users, there is no agreement about privacy requirements be- cause privacy is highly subjective and which information is considered sensitive depends on each individual. despite of what has been previously said about blurring and pixelating not being effective in providing a balance between pri- vacy and intelligibility, some users that have participated in studies about this matter feel satisfied with these filters when they are configured correctly (zhao & stasko, ; boyle et al., ; lander et al., ; neustaedter et al., ). other participants felt that there is not such a balance because as they were able to know what subjects were doing, privacy was not being suitably protected. indeed, while some of the participants showed interest in using video cameras at work and even at their homes imposing some constraints, others commented that they would not use them because video cameras are very intrusive. any- way, what can be extracted about this lack of consensus is that privacy is a very subjective topic and user’s requirements vary. considering which protection methods preserve privacy better for a fall de- tection system, a user survey was conducted by edgcomb & vahid ( ). they analysed several visual representations: blur, silhouette, oval, box, and trailing arrows. results indicated that silhouettes and blur were perceived to provide insufficient privacy, whereas an oval provides sufficient perceived privacy while still supporting fall detection accuracy of %. others have also researched about possible features of subjects’ sense of security and privacy for video surveillance systems (koshimizu et al., ; babaguchi et al., ). results showed how subjects classify viewers that monitor them using cameras in: familiar persons, unfamiliar persons and per- sons in duty. moreover, subjects expect a very familiar person to protect them in an emergency. concerning the disclosure of private information, it is af- fected by the closeness between subjects and viewers. familiarity or closeness between persons facilitates the recognition of each other. it is also curious to see how people in their s are less sensitive to privacy than those in their - s. regarding older people, most of them agree with using silhouettes as a visual representation in order to protect their privacy. furthermore, they show interest in customising the system operation. they demand control over the camera and who has access to the captured or stored information (e.g. turning a camera off, setting up the filtration level, seeing how they are being watched, and so on). however, there are others who consider that they do not require a monitoring system because they are independent enough (demiris et al., ). finally, it is necessary that users demand protection methods in order to get visual privacy protection techniques widely used in video surveillance systems. without such a demand, main actors of the security market, more interested in making a profit, will not pay attention to privacy issues because they do not have enough motivation to kick-start self-regulation (gutwirth et al., ). we consider that a strong demand of privacy-aware systems will increase research on this field also benefiting to future aal systems. either way, more discussion is needed in visual privacy, privacy requirements of users according to applications need to be collected, and more research on visual privacy protection is required. . conclusion as we have seen throughout this work, the proliferation of networks of video cameras for surveillance in public spaces and the advances in computer vision have led to a situation in which the privacy of individuals is being compro- mised. moreover, nowadays video cameras are being used more often in ambient- assisted living applications that operate in private spaces. because of this, ex- pert and intelligent systems that handle these tasks should take privacy into account by means of new tools that restore the individual’s right to privacy. in this paper, an up-to-date review of visual privacy protection methods has been provided. as far as we know, this is the first review that focuses on the protection methods and tries to give a comprehensive classification of all of them. in our classification, we have considered the following categories: intervention, blind vision, secure processing, redaction and data hiding. as it has been seen, redaction methods cover the largest part of this work and they have also been classified according to the way in which imagery data is modified: image filtering, encryption, face de-identification, object removal and visual abstraction. given that the previous methods are required for expert and intelligent sys- tems for video surveillance and ambient-assisted living, some of them have also been reviewed. such systems are developed as a framework to provide some kind of privacy protection. however, only using some of the described meth- ods is not enough to build a privacy-aware system. as it has been seen, they have to support also other mechanisms to implement privacy policies. these policies specify what data is protected, how it is protected, who has access, and a variety of other fundamental aspects concerning involved subjects, ongoing activities, captured data, and so on. despite of visual privacy protection being necessary in expert and intelligent systems for video surveillance, due to its in- creasing power of tracking, storing and indexing of our public daily activities, the use of privacy protection methods can also be justified for such systems but from an ambient-assisted living perspective. in that sense, privacy-aware smart cameras are essential in the construction of future low-cost systems for telecare applications aimed at older and disabled people. as another contribution of this work, we have also provided a discussion about important factors and current limitations of visual privacy protection. for instance, a key aspect, i.e. the correct detection of the region of interest, has been highlighted. it is a very relevant topic because it can be considered as the first building block of privacy protection and, therefore, it would deserve a self-standing review of the related research fields. we have also mentioned the problem faced by redaction methods, i.e. the privacy and intelligibility trade-off, and the necessity of expanding these studies to cover other protection methods. furthermore, we have stood out the lack of a formal model and evaluation mechanisms that would enable to make fair comparisons between different protection methods and, more importantly, it would provide guarantees about their accuracy in protecting privacy. this work provides a comprehensive picture of how to protect visual privacy, so it can be a good starting point for novel researchers in the field of expert and intelligent systems, and more specifically in intelligent monitoring systems for video surveillance and ambient-assisted living. for future research directions we consider that there are two main axis that need special attention. on the one hand, as aforementioned, a standard way to quantify and evaluate visual privacy protection is needed so as to fairly compare works in this matter. it would be better if this is included as a part of a formal model that considers more privacy related aspects. on the other hand, recognition accuracy of sensitive regions needs to be improved, so research is required in more robust computer vision algorithms for recognition, tracking and event detection. regarding other future research directions, protection methods would also deserve some consideration. although a lot of work has been done so far, novel redaction methods that provide a better balance between privacy and intelligi- bility are welcome. for instance, a textual representation of the scene, where the essential context (e.g. individuals, events, etc.) is captured, could be enough for some applications like telecare. privacy preferences are also very important. mechanisms to empower users and put them in control of their private data captured by intelligent systems should be researched. it would be appreciated to have a common way to let users specify their preferences so they can decide who, how, where and when they are watched in normal scenarios where no law infringement is involved. finally, users should be able of knowing and track- ing which systems have collected data about them in order to enable them to take legal actions in this matter just in case they need it. therefore, work in these areas may lead to obtain more effective and reliable visual privacy protec- tion methods and systems in a near future and also helps to increase the users’ acceptance of using video cameras in private spaces. acknowledgement this work has been partially supported by the spanish ministry of science and innovation under project “sistema de visión para la monitorización de la actividad de la vida diaria en el hogar” (tin - -c - ) and by the european commission under project “caring u - a study on people activity in private spaces: towards a multisensor network that meets privacy require- ments” (pief-ga- - ). josé ramón padilla lópez and alexandros andre chaaraoui acknowledge financial support by the conselleria d’educació, formació i ocupació of the generalitat valenciana (fellowship acif/ / and acif/ / respectively). the funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript. this article was originally published in josé ramón padilla-lópez, alexan- dros andre chaaraoui, francisco flórez-revuelta, visual privacy protection methods: a survey, expert systems with applications, issn: - , http: //dx.doi.org/ . /j.eswa. . . . http://dx.doi.org/ . /j.eswa. . . http://dx.doi.org/ . /j.eswa. . . references abraham, a. r., prabhavathy, k. a., & shree, d. j. ( ). a survey on video inpainting. international journal of computer applications , , – . adams, a. a., & ferryman, j. m. ( ). the future of video analytics for surveillance and its ethical implications. security journal , . doi: . /sj. . . agi, i., & gong, l. ( ). an empirical study of secure mpeg video trans- missions. in network and distributed system security, proceedings of the symposium on (pp. – ). ieee. agrawal, p., & narayanan, p. ( ). person de-identification in videos. circuits and systems for video technology, ieee transactions on, , – . augusto, j., nakashima, h., & aghajan, h. ( ). ambient intelligence and smart environments: a state of the art. in h. nakashima, h. aghajan, & j. augusto (eds.), handbook of ambient intelligence and smart environments (pp. – ). springer us. avidan, s., & butman, m. ( ). blind vision. in a. leonardis, h. bischof, & a. pinz (eds.), computer vision - eccv (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. avidan, s., elbaz, a., malkin, t., & moriarty, r. ( ). oblivious image matching. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. baaziz, n., lolo, n., padilla, o., & petngang, f. ( ). security and privacy protection for automated video surveillance. in signal processing and infor- mation technology, ieee international symposium on (pp. – ). babaguchi, n., koshimizu, t., umata, i., & toriyama, t. ( ). psychological study for designing privacy protected video surveillance system: prisurv. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. baran, i., & popović, j. ( ). automatic rigging and animation of d charac- ters. graphics, acm transactions on, . doi: . / . . barhm, m., qwasmi, n., qureshi, f., & el khatib, k. ( ). negotiating pri- vacy preferences in video surveillance systems. in k. mehrotra, c. mohan, j. oh, p. varshney, & m. ali (eds.), modern approaches in applied intelli- gence (pp. – ). springer berlin heidelberg volume of lecture notes in computer science. bertalmio, m., sapiro, g., caselles, v., & ballester, c. ( ). image inpainting. in computer graphics and interactive techniques, proceedings of the th annual conference on siggraph ’ (pp. – ). new york, ny, usa: acm press/addison-wesley publishing co. http://dx.doi.org/ . /sj. . http://dx.doi.org/ . /sj. . http://dx.doi.org/ . / . bitouk, d., kumar, n., dhillon, s., belhumeur, p., & nayar, s. k. ( ). face swapping: automatically replacing faces in photographs. graphics, acm transactions on, , : – : . borosán, p., jin, m., decarlo, d., gingold, y., & nealen, a. ( ). rigmesh: automatic rigging for part-based shape modeling and deformation. graphics, acm transactions on, , : – : . boult, t. ( ). pico: privacy through invertible cryptographic obscuration. in computer vision for interactive and intelligent environment, (pp. – ). ieee. boyle, m., edwards, c., & greenberg, s. ( ). the effects of filtered video on awareness and privacy. in computer supported cooperative work, proceedings of the acm conference on cscw ’ (pp. – ). new york, ny, usa: acm. caloyannides, m. a. ( ). society cannot function without privacy. ieee security and privacy , , – . candes, e., romberg, j., & tao, t. ( ). robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information. information theory, ieee transactions on, , – . cardinaux, f., bhowmik, d., abhayaratne, c., & hawley, m. s. ( ). video based technology for ambient assisted living: a review of the literature. jour- nal of ambient intelligence and smart environments , , – . carrillo, p., kalva, h., & magliveras, s. ( ). compression independent reversible encryption for privacy in video surveillance. eurasip journal on information security , , – . doi: . / / . article id: : . cavoukian, a. ( ). surveillance, then and now: securing privacy in public spaces . technical report information and privacy commmissioner of on- tario. chaaraoui, a. a., climent-pérez, p., & flórez-revuelta, f. ( a). an efficient approach for multi-view human action recognition based on bag-of-key-poses. in a. a. salah, j. ruiz-del solar, c. meriçli, & p.-y. oudeyer (eds.), human behavior understanding (pp. – ). springer berlin heidelberg volume of lecture notes in computer science. chaaraoui, a. a., climent-pérez, p., & flórez-revuelta, f. ( b). a review on vision techniques applied to human behaviour analysis for ambient-assisted living. expert systems with applications , , – . chan, c. s., & liu, h. ( ). gmm-qnt hybrid framework for vision-based human motion analysis. in fuzzy systems, . fuzz-ieee . ieee international conference on (pp. – ). http://dx.doi.org/ . / / chan, t., shen, j., & zhou, h.-m. ( ). total variation wavelet inpainting. journal of mathematical imaging and vision, , – . chang, y., yan, r., chen, d., & yang, j. ( ). people identification with limited labels in privacy-protected video. in multimedia and expo, ieee international conference on (pp. – ). chattopadhyay, a., & boult, t. ( ). privacycam: a privacy preserving camera using uclinux on the blackfin dsp. in computer vision and pattern recognition, . cvpr ’ . ieee conference on (pp. – ). chen, d., chang, y., yan, r., & yang, j. ( ). tools for protecting the pri- vacy of specific individuals in video. eurasip journal on advances in signal processing , , – . doi: . / / . article id: : . chen, d., chang, y., yan, r., & yang, j. ( ). protecting personal identifi- cation in video. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. chen, y., dang, g., cheng, z., & xu, k. ( ). fast capture of personalized avatar using two kinects. journal of manufacturing systems , , – . chen, y.-y., cho, c.-w., lin, s.-h., lai, h.-y., lo, y.-c., chen, s.-y., chang, y.-j., huang, w.-t., chen, c.-h., jaw, f.-s., tsang, s., & tsai, s.-t. ( ). a vision-based regression model to evaluate parkinsonian gait from monocular image sequences. expert systems with applications , , – . cheung, s.-c., paruchuri, j., & nguyen, t. ( ). managing privacy data in pervasive camera networks. in image processing, . icip . th ieee international conference on (pp. – ). cheung, s.-c., venkatesh, m., paruchuri, j., zhao, j., & nguyen, t. ( ). protecting and managing privacy information in video surveillance systems. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. cheung, s.-c. s., zhao, j., & venkatesh, m. ( ). efficient object-based video inpainting. in image processing, ieee international conference on (pp. – ). chinomi, k., nitta, n., ito, y., & babaguchi, n. ( ). prisurv: privacy protected video surveillance system using adaptive visual abstraction. in s. satoh, f. nack, & m. etoh (eds.), advances in multimedia modeling (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. cook, d. j., augusto, j. c., & jakkula, v. r. ( ). ambient intelligence: technologies, applications, and opportunities. pervasive and mobile com- puting , , – . http://dx.doi.org/ . / / cootes, t., edwards, g., & taylor, c. ( ). active appearance models. in h. burkhardt, & b. neumann (eds.), computer vision – eccv’ (pp. – ). springer berlin heidelberg volume of lecture notes in computer science. cootes, t., edwards, g., & taylor, c. ( ). active appearance models. pattern analysis and machine intelligence, ieee transactions on, , – . coudert, f. ( ). when video cameras watch and screen: privacy implications of pattern recognition technologies. computer law & security review , , – . coutaz, j., bérard, f., carraux, e., astier, w., & crowley, j. l. ( ). comedi: using computer vision to support awareness and privacy in me- diaspaces. in chi ’ extended abstracts on human factors in computing systems chi ea ’ (pp. – ). new york, ny, usa: acm. cox, i., honsinger, c., miller, m., & bloom, j. ( ). digital watermarking. journal of electronic imaging , , – . cranor, l., langheinrich, m., & marchiori, m. ( ). a p p prefer- ence exchange language . (appel . ). url: http://www.w .org/tr/ p p-preferences/ last visited: / / . criminisi, a., perez, p., & toyama, k. ( ). object removal by exemplar- based inpainting. in computer vision and pattern recognition, proceedings of ieee computer society conference on (pp. ii– –ii– vol. ). volume . criminisi, a., perez, p., & toyama, k. ( ). region filling and object removal by exemplar-based image inpainting. image processing, ieee transactions on, , – . dee, h., & velastin, s. ( ). how close are we to solving the problem of automated visual surveillance? machine vision and applications , , – . demiris, g., oliver, d., giger, j., skubic, m., & rantz, m. ( ). older adults’ privacy considerations for vision based recognition methods of eldercare ap- plications. technology and health care, , – . donoho, d. ( ). compressed sensing. information theory, ieee transac- tions on, , – . dufaux, f. ( ). video scrambling for privacy protection in video surveil- lance: recent results and validation framework. in mobile multimedia/image processing, security, and applications , proceedings of spie (pp. – ). volume . doi: . / . article id: . http://www.w .org/tr/p p-preferences/ http://www.w .org/tr/p p-preferences/ http://dx.doi.org/ . / . dufaux, f., & ebrahimi, t. ( ). scrambling for video surveillance with pri- vacy. in computer vision and pattern recognition workshop, . cvprw ’ . conference on (p. ). doi: . /cvprw. . . dufaux, f., & ebrahimi, t. ( a). h. /avc video scrambling for privacy protection. in image processing, . icip . th ieee international conference on (pp. – ). dufaux, f., & ebrahimi, t. ( b). scrambling for privacy protection in video surveillance systems. circuits and systems for video technology, ieee transactions on, , – . dufaux, f., & ebrahimi, t. ( ). a framework for the validation of privacy protection solutions in video surveillance. in multimedia and expo (icme), ieee international conference on (pp. – ). dufaux, f., ouaret, m., abdeljaoued, y., navarro, a., vergnenègre, f., & ebrahimi, t. ( ). privacy enabling technology for video surveillance. in mobile multimedia/ image processing for military and security applications, proceedings of spie (pp. – ). volume . doi: . / . article id: m. ec ( ). seniorwatch . assessment of the senior market for ict progress and developments . technical report european commission. ec ( ). overview of the european strategy in ict for ageing well . technical report european commission. edgcomb, a., & vahid, f. ( ). privacy perception and fall detection accuracy for in-home video assistive monitoring with privacy enhancements. acm sighit record , , – . efros, a., & leung, t. ( ). texture synthesis by non-parametric sampling. in computer vision, . the proceedings of the seventh ieee international conference on (pp. – ). volume . efros, a. a., & freeman, w. t. ( ). image quilting for texture synthesis and transfer. in computer graphics and interactive techniques, proceedings of the th annual conference on siggraph ’ (pp. – ). new york, ny, usa: acm. erkin, z., franz, m., guajardo, j., katzenbeisser, s., lagendijk, i., & toft, t. ( ). privacy-preserving face recognition. in i. goldberg, & m. atallah (eds.), privacy enhancing technologies (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. erturk, s. ( ). multiplication-free one-bit transform for low-complexity block-based motion estimation. signal processing letters, ieee , , – . http://dx.doi.org/ . /cvprw. . http://dx.doi.org/ . / . fan, j., luo, h., hacid, m.-s., & bertino, e. ( ). a novel approach for privacy-preserving video sharing. in information and knowledge management, proceedings of the th acm international conference on cikm ’ (pp. – ). new york, ny, usa: acm. fidaleo, d. a., nguyen, h.-a., & trivedi, m. ( ). the networked sen- sor tapestry (nest): a privacy enhanced software architecture for interactive analysis of data in video-sensor networks. in video surveillance & sensor networks, proceedings of the acm nd international workshop on vssn ’ (pp. – ). new york, ny, usa: acm. fleck, s., & strasser, w. ( ). smart camera based monitoring system and its application to assisted living. proceedings of the ieee , , – . frome, a., cheung, g., abdulkader, a., zennaro, m., wu, b., bissacco, a., adam, h., neven, h., & vincent, l. ( ). large-scale privacy protection in google street view. in computer vision, ieee th international conference on (pp. – ). gerrard, g., & thompson, r. ( ). two million cameras in the uk. cctv image, , – . ghanbari, a., & soryani, m. ( ). contour-based video inpainting. in ma- chine vision and image processing (mvip), th iranian (pp. – ). ieee. goessl, l. ( ). new japanese security camera scans million faces per second. url: http://digitaljournal.com/article/ last visited: / / . granados, m., tompkin, j., kim, k., grau, o., kautz, j., & theobalt, c. ( ). how not to be seen – object removal from videos of crowded scenes. computer graphics forum, , – . gross, r., airoldi, e., malin, b., & sweeney, l. ( a). integrating utility into face de-identification. in g. danezis, & d. martin (eds.), privacy enhancing technologies (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. gross, r., sweeney, l., cohn, j., torre, f., & baker, s. ( ). face de- identification. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. gross, r., sweeney, l., de la torre, f., & baker, s. ( b). model-based face de-identification. in computer vision and pattern recognition workshop, . cvprw ’ . conference on (pp. – ). guangzhen, l., yoshimichi, i., xiaoyi, y., naoko, n., & noboru, b. ( ). re- coverable privacy protection for video content distribution. eurasip journal on information security , , – . doi: . / / . article id: : . http://digitaljournal.com/article/ http://dx.doi.org/ . / / gutwirth, s., leenes, r., de hert, p., & poullet, y. ( ). european data protection: in good health? . springer. harvey, a. ( ). camoflash - anti-paparazzi clutch. url: http:// ahprojects.com/projects/camoflash/ last visited: / / . hayes, g., & truong, k. ( ). selective archiving: a model for privacy sen- sitive capture and access technologies. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. he, l., bleyer, m., & gelautz, m. ( ). object removal by depth- guided inpainting. in the austrian association for pattern recognition (oagm/aapr) workshop (oagm ) (pp. – ). hodgins, j. k., o’brien, j. f., & tumblin, j. ( ). perception of human mo- tion with different geometric models. visualization and computer graphics, ieee transactions on, , – . hogue, a., gill, s., & jenkin, m. ( ). automated avatar creation for d games. in future play, proceedings of the conference on future play ’ (pp. – ). new york, ny, usa: acm. hudson, s. e., & smith, i. ( ). techniques for addressing fundamental pri- vacy and disruption tradeoffs in awareness support systems. in computer sup- ported cooperative work, proceedings of the acm conference on cscw ’ (pp. – ). new york, ny, usa: acm. de la hunty, m., asthana, a., & goecke, r. ( ). linear facial expression transfer with active appearance models. in pattern recognition (icpr), th international conference on (pp. – ). igehy, h., & pereira, l. ( ). image replacement through texture synthesis. in image processing, proceedings of the international conference on (pp. – ). volume . ito, i., & kiya, h. ( ). one-time key based phase scrambling for phase- only correlation between visually protected images. eurasip journal on information security , , – . doi: . / / . article id: : . johansson, g. ( ). visual perception of biological motion and a model for its analysis. perception & psychophysics , , – . kitahara, i., kogure, k., & hagita, n. ( ). stealth vision for protecting privacy. in pattern recognition, . icpr . proceedings of the th international conference on (pp. – ). volume . kokaram, a., morris, r., fitzgerald, w., & rayner, p. ( ). interpolation of missing data in image sequences. image processing, ieee transactions on, , – . http://ahprojects.com/projects/camoflash/ http://ahprojects.com/projects/camoflash/ http://dx.doi.org/ . / / koochari, a., & soryani, m. ( ). exemplar-based video inpainting with large patches. journal of zhejiang university science c , , – . korshunov, p., araimo, c., de simone, f., velardo, c., dugelay, j., & ebrahimi, t. ( ). subjective study of privacy filters in video surveil- lance. in multimedia signal processing (mmsp), ieee th interna- tional workshop on (pp. – ). korshunov, p., & ebrahimi, t. ( ). pevid: privacy evaluation video dataset. in applications of digital image processing xxxvi, proceedings of spie (pp. – ). volume . doi: . / . article id: s. korshunov, p., & ebrahimi, t. ( ). uhd video dataset for evaluation of privacy. in sixth international workshop on quality of multimedia experience (qomex ). koshimizu, t., toriyama, t., & babaguchi, n. ( ). factors on the sense of privacy in video surveillance. in continuous archival and retrieval of personal experiences, proceedings of the rd acm workshop on carpe ’ (pp. – ). new york, ny, usa: acm. lander, k., bruce, v., & hill, h. ( ). evaluating the effectiveness of pixela- tion and blurring on masking the identity of familiar faces. applied cognitive psychology , , – . langheinrich, m. ( ). privacy by design - principles of privacy-aware ubiqui- tous systems. in ubiquitous computing, proceedings of the rd international conference on ubicomp ’ (pp. – ). springer-verlag. li, s., zhao, y., qu, b., & wang, j. ( ). image scrambling based on chaotic sequences and veginère cipher. multimedia tools and applications , , – . lin, y., wang, s., lin, q., & tang, f. ( ). face swapping under large pose variations: a d model based approach. in multimedia and expo (icme), ieee international conference on (pp. – ). lyons, m., plante, a., jehan, s., inoue, s., & akamatsu, s. ( ). avatar creation using automatic face processing. in multimedia, proceedings of the sixth acm international conference on (pp. – ). acm. ma, w., & ma, q. ( ). an effective method for removing large object. in image and signal processing (cisp), th international congress on (pp. – ). volume . macq, b., & quisquater, j.-j. ( ). cryptology for digital tv broadcasting. proceedings of the ieee , , – . http://dx.doi.org/ . / . mart́ınez-ponte, i., desurmont, x., meessen, j., & delaigle, j.-f. ( ). ro- bust human face hiding ensuring privacy. in image analysis for multimedia interactive services (wiamis), proceedings of the international workshop on. volume . masnou, s., & morel, j.-m. ( ). level lines based disocclusion. in image processing, . icip . proceedings. international conference on (pp. – ). volume . mitskog, t. f. r., & ralston, r. a. ( ). camera blocker for a device with an integrated camera that uses a thin film organic polymer. morris, m. e., adair, b., miller, k., ozanne, e., hansen, r., pearce, a. j., santamaria, n., viegas, l., long, m., & said, c. m. ( ). smart-home technologies to assist older people to live well at home. journal of aging science, . doi: . / - . . morse, b., & schwartzwald, d. ( ). isophote-based interpolation. in image processing, . icip . proceedings. international conference on (pp. – ). volume . neustaedter, c., & greenberg, s. ( ). the design of a context-aware home media space for balancing privacy and awareness. in a. dey, a. schmidt, & j. mccarthy (eds.), ubicomp : ubiquitous computing (pp. – ). springer berlin heidelberg volume of lecture notes in computer sci- ence. neustaedter, c., greenberg, s., & boyle, m. ( ). blur filtration fails to preserve privacy for home-based video conferencing. computer-human inter- action, acm transactions on, , – . newton, e., sweeney, l., & malin, b. ( ). preserving privacy by de- identifying face images. knowledge and data engineering, ieee transactions on, , – . ng, p. l., minn ang, l., & seng, k. p. ( ). privacy preserving stereoscopic vision with one-bit transform. in computer science and information tech- nology (iccsit), rd ieee international conference on (pp. – ). volume . ni, z., shi, y.-q., ansari, n., & su, w. ( ). reversible data hiding. circuits and systems for video technology, ieee transactions on, , – . o’brien, a., & mac ruairi, r. ( ). survey of assistive technology devices and applications for aging in place. in advances in human-oriented and personalized mechanisms, technologies, and services, . centric ’ . second international conference on (pp. – ). http://dx.doi.org/ . / - . olivieri, d. n., conde, i. g., & sobrino, x. a. v. ( ). eigenspace-based fall detection and activity recognition from motion templates and machine learning. expert systems with applications , , – . pande, a., & zambreno, j. ( ). securing multimedia content using joint compression and encryption. in embedded multimedia security systems (pp. – ). springer london. park, s., & kautz, h. ( ). privacy-preserving recognition of activities in daily living from multi-view silhouettes and rfid-based training. in ai in eldercare: new solutions to old problems, aaai fall symposium on. paruchuri, j., & cheung, s.-c. ( ). joint optimization of data hiding and video compression. in circuits and systems, . iscas . ieee inter- national symposium on (pp. – ). paruchuri, j. k., cheung, s.-c. s., & hail, m. w. ( ). video data hiding for managing privacy information in surveillance systems. eurasip journal on information security , , – . doi: . / / . article id: : . patel, s. n., summet, j. w., & truong, k. n. ( ). blindspot: creating capture-resistant spaces. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. patwardhan, k., sapiro, g., & bertalmio, m. ( ). video inpainting under constrained camera motion. image processing, ieee transactions on, , – . petitcolas, f., anderson, r., & kuhn, m. ( ). information hiding-a survey. proceedings of the ieee , , – . qiao, l., nahrstedt, k. et al. ( ). a new algorithm for mpeg video encryp- tion. in imaging science, systems and technology (cisst’ , proceedings of first international conference on (pp. – ). qureshi, f. ( ). object-video streams for preserving privacy in video surveil- lance. in advanced video and signal based surveillance, . avss ’ . sixth ieee international conference on (pp. – ). ra, m.-r., govindan, r., & ortega, a. ( ). p : toward privacy-preserving photo sharing. in networked systems design and implementation, proceed- ings of the th usenix conference on (pp. – ). raju, c., umadevi, g., srinathan, k., & jawahar, c. ( ). fast and secure real-time video encryption. in computer vision, graphics image processing, . icvgip ’ . sixth indian conference on (pp. – ). ieee. http://dx.doi.org/ . / / sadeghi, a.-r., schneider, t., & wehrenberg, i. ( ). efficient privacy- preserving face recognition. in d. lee, & s. hong (eds.), information, secu- rity and cryptology - icisc (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. sadimon, s., sunar, m., mohamad, d., & haron, h. ( ). computer gener- ated caricature: a survey. in cyberworlds (cw), international confer- ence on (pp. – ). saini, m., atrey, p., mehrotra, s., & kankanhalli, m. ( ). w -privacy: understanding what, when, and where inference channels in multi-camera surveillance video. multimedia tools and applications , , – . schaar, p. ( ). privacy by design. identity in the information society , , – . schiff, j., meingast, m., mulligan, d. k., sastry, s., & goldberg, k. ( ). respectful cameras: detecting visual markers in real-time to address privacy concerns. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. senior, a. ( ). privacy protection in a video surveillance system. in a. senior (ed.), protecting privacy in video surveillance (pp. – ). springer london. senior, a., & pankanti, s. ( ). privacy protection and face recognition. in s. z. li, & a. k. jain (eds.), handbook of face recognition (pp. – ). springer london. senior, a., pankanti, s., hampapur, a., brown, l., li tian, y., & ekin, a. ( ). blinkering surveillance: enabling video privacy through computer vision. technical report ibm research division. senior, a., pankanti, s., hampapur, a., brown, l., tian, y.-l., ekin, a., connell, j., shu, c. f., & lu, m. ( ). enabling video privacy through computer vision. security privacy, ieee , , – . shashank, j., kowshik, p., srinathan, k., & jawahar, c. ( ). private content based image retrieval. in computer vision and pattern recognition, . cvpr . ieee conference on (pp. – ). shashanka, m. ( ). a privacy preserving framework for gaussian mixture models. in data mining workshops (icdmw), ieee international conference on (pp. – ). shiratori, t., matsushita, y., tang, x., & kang, s. b. ( ). video completion by motion field transfer. in computer vision and pattern recognition, ieee computer society conference on (pp. – ). volume . shoaib, m., elbrandt, t., dragon, r., & ostermann, j. ( ). altcare: safe living for elderly people. in pervasive computing technologies for healthcare (pervasivehealth), th international conference on (pp. – ). sohn, h., de neve, w., & ro, y. m. ( ). privacy protection in video surveillance systems: analysis of subband-adaptive scrambling in jpeg xr. circuits and systems for video technology, ieee transactions on, , – . sohn, h., plataniotis, k., & ro, y. ( ). privacy-preserving watch list screen- ing in video surveillance system. in g. qiu, k. lam, h. kiya, x.-y. xue, c.-c. kuo, & m. lew (eds.), advances in multimedia information process- ing - pcm (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. spanos, g. a., & maples, t. b. ( ). performance study of a selective en- cryption scheme for the security of networked, real-time video. in computer communications and networks, international conference on. srinivasan, s., tu, c., regunathan, s., & sullivan, g. ( ). hd photo: a new image coding technology for digital photography. in applications of digital image processing xxx, proceedings of spie (pp. – ). volume . doi: . / . article id: a. sweeney, l. ( ). k-anonymity: a model for protecting privacy. international journal of uncertainty, fuzziness and knowledge-based systems , , – . tang, l. ( ). methods for encrypting and decrypting mpeg video data efficiently. in multimedia, proceedings of the fourth acm international con- ference on (pp. – ). acm. tansuriyavong, s., & hanaki, s.-i. ( ). privacy protection by concealing per- sons in circumstantial video image. in perceptive user interfaces, proceedings of the workshop on (pp. – ). acm. tian, y.-l., brown, l., hampapur, a., lu, m., senior, a., & shu, c.-f. ( ). ibm smart surveillance system (s ): event based video surveillance system with an open and extensible framework. machine vision and applications , , – . tong, l., dai, f., zhang, y., & li, j. ( ). prediction restricted h. /avc video scrambling for privacy protection. electronics letters , , – . tong, l., dai, f., zhang, y., li, j., & zhang, d. ( ). compressive sensing based video scrambling for privacy protection. in visual communications and image processing (vcip), ieee (pp. – ). turk, m., & pentland, a. ( ). face recognition using eigenfaces. in com- puter vision and pattern recognition, . proceedings cvpr ’ ., ieee computer society conference on (pp. – ). http://dx.doi.org/ . / . upmanyu, m., namboodiri, a., srinathan, k., & jawahar, c. ( ). efficient privacy preserving video surveillance. in computer vision, ieee th international conference on (pp. – ). vijay venkatesh, m., cheung, s.-c. s., & zhao, j. ( ). efficient object-based video inpainting. pattern recognition letters , , – . wagstaff, j. ( ). using bluetooth to disable camera phones. url: http://www.loosewireblog.com/ / /using_bluetooth.html last vis- ited: / / . wallace, e., & diffley, c. ( ). cctv control rooms ergonomics . technical report / police scientific development branch (psdb) uk home office. wexler, y., shechtman, e., & irani, m. ( ). space-time completion of video. pattern analysis and machine intelligence, ieee transactions on, , – . whyte, o., sivic, j., & zisserman, a. ( ). get out of my picture! internet- based inpainting. in proceedings of the th british machine vision confer- ence, london. williams, a., xie, d., ou, s., grupen, r., hanson, a., & riseman, e. ( ). distributed smart cameras for aging in place. technical report dtic doc- umen. winkler, t., & rinner, b. ( a). securing embedded smart cameras with trusted computing. eurasip journal on wireless communications and net- working , , – . doi: . / / . article id: : . winkler, t., & rinner, b. ( b). trustcam: security and privacy-protection for an embedded smart camera based on trusted computing. in advanced video and signal based surveillance (avss), seventh ieee interna- tional conference on (pp. – ). winkler, t., & rinner, b. ( ). security and privacy protection in visual sensor networks: a survey. acm computing surveys , , – . wu, m. ( ). multimedia data hiding . ph.d. thesis princeton university. xiangdong, l., junxing, z., jinhai, z., & xiqin, h. ( ). image scrambling algorithm based on chaos theory and sorting transformation. ijcsns inter- national journal of computer science and network security , , – . yabuta, k., kitazawa, h., & tanaka, t. ( ). a new concept of security camera monitoring with privacy protection by masking moving objects. in y.-s. ho, & h. kim (eds.), advances in multimedia information processing - pcm (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. http://www.loosewireblog.com/ / /using_bluetooth.html http://dx.doi.org/ . / / yabuta, k., kitazawa, h., & tanaka, t. ( ). a new concept of real-time se- curity camera monitoring with privacy protection by masking moving objects. in proceedings of spie (pp. – ). volume . yang, m., bourbakis, n., & li, s. ( ). data-image-video encryption. poten- tials, ieee , , – . yu, x., & babaguchi, n. ( ). privacy preserving: hiding a face in a face. in y. yagi, s. kang, i. kweon, & h. zha (eds.), computer vision - accv (pp. – ). springer berlin / heidelberg volume of lecture notes in computer science. yu, x., chinomi, k., koshimizu, t., nitta, n., ito, y., & babaguchi, n. ( ). privacy protecting visual processing for secure video surveillance. in image processing, . icip . th ieee international conference on (pp. – ). zeng, w., & lei, s. ( ). efficient frequency domain selective scrambling of digital video. multimedia, ieee transactions on, , – . zhang, c., rui, y., & wei he, l. ( ). light weight background blurring for video conferencing applications. in image processing, ieee interna- tional conference on (pp. – ). zhang, c., tian, y., & capezuti, e. ( ). privacy preserving automatic fall detection for elderly using rgbd cameras. in k. miesenberger, a. karshmer, p. penaz, & w. zagler (eds.), computers helping people with special needs (pp. – ). springer berlin heidelberg volume of lecture notes in computer science. zhang, w., cheung, s., & chen, m. ( a). hiding privacy information in video surveillance system. in image processing, . icip , ieee in- ternational conference on (pp. ii– – ). volume . zhang, y., xiao, j., & shah, m. ( b). motion layer based object removal in videos. in application of computer vision, . wacv/motions ’ volume . seventh ieee workshops on (pp. – ). volume . zhao, q. a., & stasko, j. t. ( ). the awareness-privacy tradeoff in video sup- ported informal awareness: a study of image-filtering based techniques . tech- nical report georgia institute of technology. gvu technical report;git- gvu- - . zivkovic, z. ( ). improved adaptive gaussian mixture model for background subtraction. in pattern recognition, . icpr . proceedings of the th international conference on (pp. – ). volume . introduction related studies visual privacy protection protection methods intervention blind vision secure processing redaction image filtering encryption of videos and images face de-identification object / people removal visual abstraction / object replacement data hiding based methods privacy-aware intelligent monitoring systems discussion recognition of the region of interest privacy and intelligibility trade-off real-time applications privacy model and evaluation user studies about privacy requirements conclusion wp-p m- .ebi.ac.uk params is empty sys_ exception wp-p m- .ebi.ac.uk no params is empty exception params is empty / / - : : if (typeof jquery === "undefined") document.write('[script type="text/javascript" src="/corehtml/pmc/jig/ . . /js/jig.min.js"][/script]'.replace(/\[/g,string.fromcharcode( )).replace(/\]/g,string.fromcharcode( ))); // // // window.name="mainwindow"; .pmc-wm {background:transparent repeat-y top left;background-image:url(/corehtml/pmc/pmcgifs/wm-nobrand.png);background-size: auto, contain} .print-view{display:block} page not available reason: the web page address (url) that you used may be incorrect. message id: (wp-p m- .ebi.ac.uk) time: / / : : if you need further help, please send an email to pmc. include the information from the box above in your message. otherwise, click on one of the following links to continue using pmc: search the complete pmc archive. browse the contents of a specific journal in pmc. find a specific article by its citation (journal, date, volume, first page, author or article title). http://europepmc.org/abstract/med/ a reduced data set method for support vector regression a reduced data set method for support vector regression horng-lin shieh *, cheng-chien kuo department of electrical engineering, saint john’s university, taipei, taiwan a r t i c l e i n f o keywords: support vector regression outlier fuzzy clustering robust fuzzy c-means a b s t r a c t support vector regression (svr) has been very successful in pattern recognition, text categorization, and function approximation. the theory of svr is based on the idea of structural risk minimization. in real application systems, data domain often suffers from noise and outliers. when there is noise and/or out- liers exist in sampling data, the svr may try to fit those improper data, and obtained systems may have the phenomenon of overfitting. in addition, the memory space for storing the kernel matrix of svr will be increment with o(n ), where n is the number of training data. hence, for a large training data set, the kernel matrix cannot be saved in the memory. in this paper, a reduced support vector regression is pro- posed for nonlinear function approximation problems with noise and outliers. the core idea of this approach is to adopt fuzzy clustering and a robust fuzzy c-means (rfcm) algorithm to reduce the com- putational time of svr and greatly mitigates the influence of data noise and outliers. crown copyright � published by elsevier ltd. all rights reserved. . introduction the theory of support vector machines (svm) developed by vapnik ( ) in is gaining in popularity due to its many attractive features. the svm is based on the idea of structural risk minimization (srm) and has been shown to be superior to tradi- tional empirical risk minimization (erm) principles employed by conventional neural networks (gunn, ). svm has been suc- cessfully applied to a number of applications, such as classification, time predictions, pattern recognition, and regression (burges, ; jair, xiaoou, wen, & kang, ; kamruzzaman & begg, ; kumar, kulkarni, jayaraman, & kulkarni, ; lijuan, ; wong & hsu, ; zhou, zhang, & jiao, ). in many intelligent systems, svm has been shown to provide higher perfor- mance than traditional learning machines, and has thus been adopted as a tool for solving classification issues (lin & wang, ). over the past few years, a lot of researchers of neural net- works and machine learning fields are attracted to devoting them- selves to research on svm (wang & xu, ). the svm is systematic and properly motivated by the statistical learning theory (vapnik, ). training of the svm involves opti- mization of a convex cost function and globally minimizes to com- plete the learning process (campbell, ). in addition, svm can handle large input, and can automatically identify a small subset consisting of informative points, namely support vectors (gustavo et al., ). the svm can also be applied to regression problems by the introduction of an alternative loss function (gunn, ). such approaches are often called support vector regression (svr). svm maps the input data into a high-dimensional feature space, and searches a separate hyperplane that maximizes the margin be- tween two classes. svm adopts quadratic programming (qp) to maximize the margin as computing tasks become very challenging when the number of data is beyond a few thousand (hu & song, ). for example, in fig. , there are sampling data gener- ated from a sin wave with gaussian noise n( , . ). the svr algo- rithm is adopted to construct this function. the entries of the kernel matrix of svr are floating-points numbers, and each float- ing-point number requires bytes for storing. therefore, the total memory required is � � = , bytes. the svr algo- rithm is performed on a pentium , . ghz with mb of mem- ory running windows xp. the total execution time of the simulation is s (above h). this execution time is very long and the memory requirements are very large for real applications of science. osuna, freund, and girosi ( ) proposed a generalized decomposition strategy for the standard svm, in which the original qp problem is replaced by a series of smaller sub-problems, which are proved able to converge to a global optimum point. however, it is well known that the decomposition process relies heavily on the selection of a good working set of the data, which normally starts with a random subset (hu & song, ). lee and huang ( ) proposed to restrict the number of support vectors by solving re- duced support vector machines (rsvm). the main characteristic of this method is to reduce the matrix from l � l to l � m, where m is the size of a randomly selected subset of training data that are considered as candidates of support vectors. the smaller matrix - /$ - see front matter crown copyright � published by elsevier ltd. all rights reserved. doi: . /j.eswa. . . * corresponding author. e-mail address: shieh@mail.sju.edu.tw (h.-l. shieh). expert systems with applications ( ) – contents lists available at sciencedirect expert systems with applications j o u r n a l h o m e p a g e : w w w . e l s e v i e r . c o m / l o c a t e / e s w a http://dx.doi.org/ . /j.eswa. . . mailto:shieh@mail.sju.edu.tw http://www.sciencedirect.com/science/journal/ http://www.elsevier.com/locate/eswa can easily be stored in memory, and then optimization algorithms, such as the newton method, can be applied (lin & lin, ). how- ever, as shown by lin and lin ( ), numerical experiments show that the accuracy of rsvm is usually lower than that of svm. because support vectors can state the distributional features of all data, according to the characteristics of svm, removing trivial data from the whole training set will not greatly affect the out- come, but will effectively increase the training process (wang & xu, ). a reduced set method based on the measurement of similarities between samples is developed by wang and xu ( ). in this paper, the samples similar to some data points will be discarded under a pre-established similarity threshold. in other words, these samples are so similar to the special data point that their influence on the prediction function can be ignored. accord- ing to this method, a large number of training vectors are dis- carded, and then a faster svm training can be obtained without compromising the generalization capability of svm. however, like the k-means clustering algorithm, the disadvantage of this algo- rithm is that the number of clusters must be predetermined, but in some real applications, there is no information to predefine the number of the clusters. in real applications, data is bound to have noise and outliers, and algorithms utilized in engineering and scientific applications must be robust in order to process these data. in system modeling with noise and/or outliers existing in the sampling data, the system models may try to fit those improper data, and the output may have the phenomenon of overfitting (chung, su, & hsiao, ; shieh, yang, chang, & jeng, ). svr has been shown to have excellent performance for both the e-insensitive and huber’s ro- bust function for matching the correct type of noise in an applica- tion of time series prediction (mukherjee, osuna, & girosi, ). however, in this svr approach, outliers may possibly be taken as support vectors, and such an inclusion of outliers in support vec- tors may lead to serious overfitting phenomena (chung, ). in this paper, in order to overcome the above problems, a robust fuzzy clustering method is proposed to greatly mitigate the influ- ence of noise and outliers in sampling data, and then the svr method is used to construct the system models. three experiments are illustrated, and their results have shown the proposed ap- proach has better performance and less execution time than the original svr method in various kinds of data domains with data noise and outliers. . support vector regression the model of learning from examples can be considered a gen- eral statistic framework of minimizing expected loss using sam- pling data. suppose there are n random independent identically distributed (i.i.d.) data (x , y ), (x , y ), . . ., (xn, yn), where xi rd, yi r, i = , , . . ., n drawn according to the uniform probability dis- tribution function p(x, y) = p(x)p(y—x). given a set of functions f(x, a), a k, where k is a parameter set, from which the goal of the learning process is to choose a function f(x, a ) that can obtain the best relationship between input and output pairs. consider a measure of the loss l(y, f(x, a )) between the output y of the sam- pling data to a given input x, and the response f(x, a ), provided by the learning machine. in order to obtain f(x, a ), one has to min- imize the expected risk functional rðaÞ¼ z lðy; fðx; a ÞÞdpðx; yÞ; ð Þ a common choice for the loss function is l -norm; i.e., l(e) = e = (y � f(x, a )) . however, because p(x, y) is unknown, r(a) cannot be directly evaluated from eq. ( ). in general, the ex- pected risk function is replaced by the empirical risk functional rempðaÞ¼ n xn i¼ lðy; fðx; a ÞÞ: ð Þ there is no probability distribution in eq. ( ). however, in real application systems, data domains often suffer from noise and out- liers. when there is noise and/or outliers exist in sampling data, eq. ( ) may try to fit those improper data and obtained systems may have the phenomenon of overfitting. let the sampling data be represented as {(xi, yi)jxi rd, yi {� , }}, i = , , . . . n. in the svr method, the regression func- tion is approximated by the following function as: f ¼ xn i¼ wiuðxiÞþ b; ð Þ where fuðxiÞg n i¼ are the features of inputs, fwig n i¼ and b are coeffi- cients. the coefficients are estimated by minimizing the regularized risk function (wang & xu, ) rðcÞ¼ c n xn i¼ lðy; fÞþ kwk ; ð Þ where l(y, f) adopt the e-insensitive loss function, and is defined as follows: lðy; fÞ¼ jy � f j� e; jy � f j p e; ; otherwise � ð Þ and e p is a predefined parameter. in eq. ( ), the second term, kwk , is used for the flatness mea- surement of function ( ), and c is a regular constant determining the tradeoff between the training error and the model flatness. svr introduces slack variables n, n* and leads eq. ( ) to the follow- ing constrained function (wang & xu, ): minimize rðw; n�Þ¼ c� xn i¼ ðni þ n � i Þþ kwk ; ð Þ subject to wuðxiÞþ b � yi e þ n � i ; ð Þ yi � wuðxiÞ� bi e þ ni; n; n� p ; where n, n* are slack variables representing upper and lower con- straints on the outputs of the system. thus, function ( ) becomes the explicit form: fðx; ai; a�i Þ¼ xn i¼ wiuðxiÞþ b ¼ xn i¼ ðai � a�i ÞuðxiÞ t uðxiÞþ b ð Þ - - . - . - . - . . . . . x y fig. . sin wave with noise n( , . ). h.-l. shieh, c.-c. kuo / expert systems with applications ( ) – https://isiarticles.com/article/ introduction fuzzy data envelopment analysis: a discrete approach majid zerafat angiz l. ali emrouznejad† adli mustafa department of mathematics, islamic azad university, firouzkooh, iran aston business school, aston university, birmingham, uk school of mathematical sciences, universiti sains malaysia, penang, malaysia abstract data envelopment analysis (dea) as introduced by charnes et al [ ] is a linear programming technique that has widely been used to evaluate the relative efficiency of a set of homogenous decision making units (dmus). in many real applications, the input-output variables cannot be precisely measured. this is particularly important in assessing efficiency of dmus using dea, since the efficiency score of inefficient dmus are very sensitive to possible data errors. hence, several approaches have been proposed to deal with imprecise data. perhaps the most popular fuzzy dea model is based on -cut. one drawback of the -cut approach is that it cannot include all information about uncertainty. this paper aims to introduce an alternative linear programming model that can include some uncertainty information from the intervals within the α-cut approach. we introduce the concept of “local α-level” to develop a multi- objective linear programming to measure the efficiency of dmus under uncertainty. an example is given to illustrate the use of this method. keywords: fuzzy data envelopment analysis; interval data; local α-level, multi objective programming, decision making unit. . introduction data envelopment analysis (dea) initially was proposed by charnes et al. [ ] is a non-parametric linear programming technique for measuring the relative efficiency of a set of homogeneous decision making units (dmus), with the common set of inputs and outputs. examples of dea include the efficiency of hospitals in providing their services [ ] measurement efficiency of health † corresponding author: ali emrouznejad, operations and information management group, aston business school, aston university, birmingham, united kingdom a.emrouznejad@aston.ac.uk file:///d:/hd/rh -pb/subm/-% rj% -% zarafat-fuzzydea-discreteapproach/-rj- - - -resubmit% in% fss/a.emrouznejad@aston.ac.uk centers [ ] manufacturing efficiency [ , ] productivity of oecd countries [ , , ]. for some computational calculation of dea methods see emrouznejad [ ] and for a recent theoretical survey and full list of applications of dea see cook and seiford [ ] and emrouznejad et al. [ ]. traditionally, all input/output values of dmus are crisp data, hence, most of the previous studies dealt with precise data. theoretically, dea measures the efficiency of each dmu by finding the distance of the dmu to the best practice; therefore, the efficiency scores are very sensitive to the data. if there is an outlier, then the efficiency scores of many dmus may change substantially. therefore, a key to the success of the dea is to measure all inputs outputs accurately. however, in real application of production process many complicated factors are involved that makes difficult to measure inputs and outputs precisely. this makes a case where we need to measure the efficiency of dmus with inexact values or interval data. several approaches have been developed to deal with fuzzy data in dea. sengupta [ ] applied principle of fuzzy set theory to introduce fuzziness in the objective function and the right-hand side vector of the conventional dea model [ ]. guo and tanaka [ ] used the ranking method and introduced a bi-level programming model. lertworasirikul [ ] developed the method in which first, inputs and outputs were defazified and then the model was solved using the α- cut approach. their method is simple but the uncertainty in inputs and outputs is effectively ignored. some other approaches based on α-cut can be found in [ , , ] and the methods based on the interval efficiency are seen in [ , ]. lertworasirikula [ ] considered each constraint in the dea as a fuzzy event; hence he transferred fuzzy dea model to possibility linear programming problem. lertworasirikul et al. [ ] further developed a fuzzy bcc model where the possibility and credibility approaches are provided and compared with an α -cut level based approach for solving the fdea models. using the possibility approach, they revealed the relationship between the primal and dual models of fuzzy bcc. using the credibility approach they showed how the efficiency value for each dmu can be obtained as a representative of its possible range. a different approach based on possibility programming was developed in [ , ]. inuiguchi and tanino [ ] applied the extension principle to define fuzzy efficiency score using dea. in this paper we develop an alternative method which is able to provide fuzzy efficiency measures for dmus with fuzzy observations. we introduce the concept of local α-level which can include more information from uncertain data into the model. use of local α-level in the proposed dea model enables us to capture as much information as possible from the uncertain dmu while in the standard α-cut some fuzzy characteristics of dmus are ignored [ - ]. the rest of this paper is organized as follows. dea and fuzzy dea are defined in section . using concept of local α-level, an alternative fuzzy dea is proposed in section . further discussion is given in section . this is followed by a numerical example and comparison with other models in section . conclusion is given in section . . dea and fuzzy dea data envelopment analysis (dea) is a non-parametric technique for measuring the relative efficiency of the decision making units (dmus) that have homogenous inputs and outputs. dea applies linear programming techniques to the observed inputs /outputs of dmus by constructing an efficient production frontier based on the best practices. each dmu's efficiency is then measured relative to its distance to this frontier [ ]. consider a set of n dmus, in which ( , ,..., ) ij x i m and ( , ,..., ) rj y r s are inputs and outputs of j dmu (j= , ,…,n). the standard form of ccr model for assessing p dmu is written as: ( ) max . . , ,              s r rp r m i ip i s m r rj i ij r i r i u y s t v x u y v x u v r i the above model can only be used for cases where the data are precisely measured. fuzzy dea is a powerful tool for evaluating the performance of dmus with imprecise data (or interval data). fuzzy input-output variables can be introduced to dea in the following fuzzy linear programming model. ( ) max . . , , s r rp r m i ip i s m r rj i ij r i r i u y s t v x u y v x j u v r i               where “~”indicates the fuzziness. rj y and ij x are fuzzy inputs and fuzzy outputs, respectively.  is a non-archimedean small positive number. saati et al. [ ] suggested a different ccr model for assessment of fuzzy data by transferring the standard ccr model to a possibilistic programming problem. their basic idea is using α-cut approach to transform the fuzzy ccr model ( ) into a crisp linear programming problem such as the standard dea model. their proposed approach assumes that the solution lies in the interval and the result for each dmu is an interval efficiency score rather than a crisp efficiency score. the main drawback in this approach is that their model can not retain the uncertainty information completely since it is based on simple α-cut approach. in other words, the fuzzy numbers are simply converted to intervals using the same membership numbers in the entire of interval. . an alternative fuzzy dea model under uncertainty the proposed approach in this paper is based on alternative definition of the membership functions of the coefficients, i.e. input output variables. assume ( )f  is the set of fuzzy number. we define “local α-level” as follows. definition : the crisp set of elements that belong to the fuzzy set m with degree of at least h and less than h  , in which h h   , is called the “local α- level” set and it is represented as follows:          | ( )h h hmm x x x ( ) one advantage of using a local α-level is that the corresponding point on the membership function retained the whole of uncertainty in process of solving the problem. we define the membership function .  m n based on  m and  n [ ]. definition : if , ( ) m n f are two fuzzy numbers that ( ), ( ) n m x y  are continuous membership functions, then  . . ( ) sup min ( ), ( )     m n m n z x y z x y ( ) obviously, if the bounded variable v is considered as a trapezoidal fuzzy number, the following corollary is obtained. corollary : the product of bounded variable v in fuzzy number m is defined as  . ( ) ( ), . ,   v m mz x z v x x x ( ) in most dea applications for the sake of computational efficiency and ease of data acquisition, trapezoidal or triangular membership functions are often used. figure a: triangular membership function figure b: concept of local α-level figure a shows a triangular “fuzzy number” and figure b shows the local α-level for a fuzzy number which is corresponding with the figure a. as seen, each α-cut acts in a local determined domain. for instance, the corresponding domain of l is [ , ] [ , ]l l l lx y x y    . the range of each local α-level in the figure b has shown as the horizontal bold lines. theorem . let ),,( ~ ulm xxxa  be a symmetric triangular fuzzy number such that ),..., , ( ni i  are local α-levels, thus, nn xxxx ,...,,...,, make a partition on the interval of fuzzy number a ~ . then, for each of the data point ],[ n xxx  we have xxxxxx ii n i a m ii n i a    )()( ~ ~  ( ) proof. to prove this theorem, we rearrange ( ) to       ll h  h h h  l l x y  l l x y  ))(( ~   xxxxx i m ii n i a  ( ) suppose that ))(( ~ xxxxxb i m ii n i a     two cases are considered case . m kk xxxx   in this case, b can be written in the following form. ))())((( )())((())())((( ~ ~ ~ xxxxx xxxxxxxxxxb i m n ni iia i n ki i m iaii m i k i a           ( ) since a ~ is symmetric, therefore nlxx lnala )()( ) ( ~~   . thus, the right side of fuzzy number a ~ is divided to two areas as follows: ))())((( ))())((())())((( )) ( ()) ( ()) ( ( ~ )) ( ( )) ( ()) ( ( ~ ~ xxxxx xxxxxxxxxx in m n ki inina in m k i ininai m n ni iia               ( ) combining ( ) and ( ) we obtain ))]()(())())[((( ))]()(())())[((( ) ( ()) ( ( ~ ) ( () ( ( ~ xxxxxxxxx xxxxxxxxxb in m ini n ki i m ia in m inii m i k i a            ( ) in equation ( ) we have ) )(( ~ xxxb ii n ki a     since xx i  for ki  , therefore, . b case . nxxxx kk m    similarly, it can be proved that . b ■ hence, xxx ii n i ax   )(min ~ in this method is obtained the maximal degree of membership function which does not depend on number of partitions. corollary . suppose ,, xx i )(~ ia x satisfy the conditions of theorem then xxx ii n i ax   )(min ~ is unique. using the concept of local α-level in model ( ) the following fuzzy linear programming is proposed for assessing p dmu . ( )                                 min ( ) , min ( ) , max . . , , , , ij rj ij x ijh i ij i ijh h rj y rjk r rj r rjk k s r rp r m i ip i s m r rj i ij r i l u ij ij ij l u rj rj rj r i z x v x v x i j z y u y u y r j z u y s t v x u y v x j x x x i j y y y r j u v r i as seen in figure b, corresponding to each α, a set of subintervals is assigned to each fuzzy number. theorem . the above model is an extension to proposed model of kao and liu [ ], if we remove the two minimization distance functions from the objective of the model, the largest value of optimal solution in each α-level is the same as those obtained in kao and liu [ ]. proof. consider the following model: ( ) max . . ( ) ( ) ( ) ( ) , , s r rp r m i ip i s m r rj i ij r i m l m u rj rj rj rj rj m l m u ij ij ij ij ij r i u y s t v x u y v x y y y y y x x x x x u v r i                                 for α= , consider n dmus with m inputs and r outputs as follows: u l ij ij io ip l u rj rj ro rp x x j o x x y y j o y y       it is clear that for o dmu each dominated dmu (a dmu with higher level of inputs and lower level of outputs) will not be more efficient. so optimal value of mathematical programming ( ) is equals to efficiency of p dmu . this is also valid for any arbitrary . in model ( ), ijhx corresponds to the length of input value of xij located in the intersection of h  and sides of the corresponding triangular membership function. similarly rjky corresponds to the length of output value of yij. consider the following variable substitutions. ˆ ijx = i ijv x , ˆ rjy = , ,r rju y i r j hence, model ( ) is concluded. ( )                                 ˆmin ( ) , ˆmin ( ) , ˆmax . . ˆ ˆ ˆ ˆ , ˆ , , , ij rj ij x ijh ij i ijh h rj y rjk rj r rjk k s p rp r m ip i s m rj ij r i l u i ij ij i ij l u r rj rj r rj r i z x x v x i j z y y u y r j z y s t x y x j v x x v x i j u y y u y r j u v r i assume that ˆ , ( , , ) ijh ij i ijhx x v x i j h    , so ( , , )ijh ijh ijhx x x i j h      and ˆ rjk rj r rjky y u y   , so ( , , )rjk rjk rjky y y r j k      . applying these substitutions model ( ) may be solved using the following multi-objective programming. ( )                                         min ( )( ) , min ( )( ) , ˆmax . . ˆ ˆ ˆ ˆ ( ) , , ˆ -( ) , , ij rj ij x ijh ijh ijh h rj y rjk rjk rjk k s p rp r m ip i s m rj ij r i ij i ijh ijh ijh rj r rjk rjk rjk l i ij z x x x i j z y y y r j z y s t x y x j x v x x x i j h y u y y y r j k v x          ˆ , ˆ , , , u ij i ij l u r rj rj r rj r i x v x i j u y y u y r j u v r i the model ( ) is a multi-objective, hence, it can be solved by archimedean goal programming model ( ) as follows: ( )                                               min . . ( )( ) , ( )( ) , ˆ ˆ ˆ ˆ ˆ ( ij rj m n s n ij ij rj rj p p i j r j x ijh ijh ijh ij ij h y rjk rjk rjk rj rj k s rp p p r m ip i s m rj ij r i ij i ijh ijh w d w d w d s t x x x d t i j y y y d t r j y d t x y x j x v x x                    ) , , ˆ -( ) , , ˆ , ˆ , , , ijh rj r rjk rjk rjk l u i ij ij i ij l u r rj rj r rj r i x i j h y u y y y r j k v x x v x i j u y y u y r j u v r i in model ( ), the w’s in the objective function are positive penalty weights and d’s measure the over-achievement and under-achievement from the target point t, i.e. ijt , rjt , pt . . discussion in figure a, consider the local α-level  . the membership values corresponding to interval [ , ]n n are approximated by  . for instance, the membership value related to x is  instead of ( )x that is real membership value. furthermore, assume that we include another local α-level  (see figure b), it is seen that the membership function corresponding to x is now  instead of ( )x . it is clear that   ( )x and  is a better approximation than  for ( )x . according to theorems , if the objective functions minimizing in ( ) are deleted from the model, the optimal solution for inputs and outputs will be arisen at its endpoints of interval of fuzzy numbers. furthermore, if the objective function maximizing in ( ) is eliminated, theorem is adopted and its optimal solutions are fuzzy number. figure illustrates the above mentioned concept for evaluating p dmu . in this figure, the interior arrows represent the optimal solution when the objective function of maximizing is absent in ( ) and the arrows located under fuzzy numbers construct the optimal solution ( ) when only objective function of maximizing is present. interaction between the objective functions of maximizing and minimizing in ( ) cause the fuzzy optimal solution. in the methods based on α-level approach, all fuzzy numbers are dealt with the same level. consider the example given in kao and liu [ ] evaluating four dmus with single input and single output. in total, lps should be solved using the α-cuts of the efficiency scores for eleven α values, they assumed that all fuzzy inputs and outputs are in the same α-level. in the proposed method different level of α’s are considered for inputs and outputs, hence we need to solve × × lps. figure b: a presentation of local two α-levels figure a: a presentation of local one α-level ij x~ ip x~ rp y~  x n n n n  ( )x  x ( )x . an application and comparison with other methods consider dmus with two triangular fuzzy inputs and triangular fuzzy outputs. table ( ) shows the data which are also used in guo and tanaka [ ]. table . data for numerical example dmu variable d d d d d i ( . , . , . ) ( . , . , . ) ( . , . , . ) ( . , . , . ) ( . , . , . ) i ( . , . , . ) ( . , . , . ) ( . , . , . ) ( . , . , . ) ( . , . , . ) o ( . , . , . ) ( . , . , . ) ( . , . , . ) ( . , . , .. ) ( . , . , . ) o ( . , . , . ) ( . , . , . ) ( . , . , . ) ( . , . , . ) ( . , . , . ) source: guo and tanaka ( ) fuzzy efficiencies of dmus using standard fuzzy dea model ( ) and with different α value solved by the method suggested in [ ] is reported in table ( ). table .the results of the model suggested in [ ] dmu α d d d d d . . . . . . . . . . . . . . . . . . . . . . using the kao and liu’s approach, the cut  of the efficiency scores for four  values is presented in table ( ). table . table of efficiency using kao & liu fuzzy dea model dmu α d d d d d ( . , . ) ( . , . ) ( . , . ) ( . , . ) ( . , . ) . ( . , . ) ( . , . ) ( , . ) ( , . ) ( . , . ) . ( . , . ) ( . , . ) ( . , . ) ( . , . ) ( . , . ) . . . . . the results of the possibility approach introduced in [ ] have given in table ( ). table . table of efficiency using lertworasirikul’s possibilistic model [ ] dmu α d d d d d . . . . . . . . . . . . . . . . . . . . . . . . . . . . although the lertworasirikul’s possibilistic model is not directly comparable with the new suggested method, the result of table ( ) is aggregated measure in a linear approach and an optimal cut  is obtained for each fuzzy numbesults to make it comparable. fuzzy efficiencies of dmus using the proposed model and with different α values are reported in table ( ). table . table of efficiency using proposed fuzzy dea model α d d d d d , . , . , . , , , . , . . . . . , . , . , . , . , . . . . . , . , . , . . . . . . . . . . the results shown in the first row of table are more accurate than the results suggested in [ ], as seen in theorem , kao and liu [ ] model obtained the results at endpoints of the intervals only. figure shows the structure of the discrete triangular number concerned to input of dmu . as it can be seen we included local α-levels for calculation of efficiency in this example.  h  , . , . , . , . , . and . rj y~ figur e : opti mal solut ions at end point inter vals . . . . . . . . . . . . figu re . disc rete fuzz y num ber h x . . . . . it should be noted that increasing the numbers of local α-levels will result more precise measure of efficiency. for example when measuring efficiency of dmu if α-levels are [ , . , . , ] we obtain efficiency score of “ . ”, while if α-levels are [ , . , . , . , . , ] we obtain efficiency score of . which is a more accurate estimation of efficiency. comparing these results with table ( ) it is clear that our proposed approach is given a better estimation of the efficiency scores when data are in the form of interval values. one drawback of the proposed model is longer computational calculation; however this is not a major issue with development of high-speed computers. . conclusion in evaluating dmus in fuzzy dea there are four traditional approaches; the fuzzy ranking approach, the defuzzification approach, the tolerance approach and the α-cut based approach. each of these methods has its advantages and drawbacks in the way they treat uncertain data in dea models. perhaps due to its simplicity, the α-cut based approach is frequently used by dea scholars. as result of simplicity in this approach we will lose a lot of uncertainty information. this paper proposed an alternative fuzzy dea technique for measuring efficiency of decision making units under fuzzy environment using local α-level concept and linear programming problem. the numerical example showed a better estimation of efficiency when using the proposed model. the final model presented is a multi-objective programming, its transformation to a linear programming and developing an efficient algorithm for large scale problems are subjects for future development. references [ ] i. alp, possibilistic data envelopment analysis, mathematical and computational applications. ( ) : - . [ ] b. asady, a. zendehnam, ranking fuzzy numbers by distance minimization, applied mathematical modeling. ( ): - . [ ] a. charnes, w.w. cooper, e. rhodes, measuring the efficiency of decision making units, european journal of operation research. ( ): - . [ ] w. d. cook, l. seiford, data envelopment analysis (dea) – thirty years on, european journal of operational research, ( ) in press: doi: . /j.ejor. . . [ ] a. emrouznejad, an alternative dea measure: a case of oced countries, applied economic letters. ( ): - . [ ] a. emrouznejad, e. thanassoulis, a mathematical model for dynamic efficiency using data envelopment analysis, journal of applied mathematics and computation. ( ) : - . [ ] a. emrouznejad, e. thanassoulis, measurement of productivity index with dynamic dea, international journal of operational research ( ), in press. [ ] a. emrouznejad, measurement efficiency and productivity in sas/or, journal of computers and operation research. ( ) : - . [ ] a. emrouznejad, b. parker, g. tavares, evaluation of research in efficiency and productivity: a survey and analysis of the first years of scholarly literature in dea, socio-economic planning sciences. ( ) : - . [ ] t. entani, y. maeda, h. tanaka, dual models of interval dea and its extension to interval data, european journal of operational research. ( ) : - . [ ] r. e. esteuer, multiple criteria optimizing: theory, computation and application, john wiley & sons, inc . [ ] k. field, a. emrouznejad, measuring the performance of neonatal care units in scotland, journal of medical systems. ( ) : - . [ ] p. guo, h. tanaka, fuzzy dea: a perceptual evaluation method." fuzzy sets and systems. ( ) : - . [ ] j. l. hougaard, a simple approximation of productivity scores of fuzzy production plans, fuzzy sets and systems. ( ) : - . [ ] m. inuiguchi and t. tanino, data envelopment analysis with fuzzy input-output data, y. haimes and r. e. steuer: research and practice in multiple criteria decision making, - , springer ( ). [ ] c. kao, s. t. liu, fuzzy efficiency measures in data envelopment analysis, fuzzy sets and systems. ( ): – . [ ] j. m. kirigia, a. emrouznejad, r. g. vaz, h. bastiene, j. padayachy, a comparative assessment of performance and productivity of health centres in seychelles, international journal of productivity & performance management. ( ) . [ ] t. leon, v. liern, j. l. ruiz, i. sirvent, a fuzzy mathematical programming approach to the assessment of efficiency with dea models, fuzzy sets and systems. ( ): - . [ ] s. lertworasirikul, s. c. fang, a. jeffrey, j. a. joines, h. l. w. nuttle, fuzzy data envelopment analysis (dea): a possibility approach, fuzzy sets and systems. ( ): - . [ ] s. lertworasirikul, fuzzy data envelopment analysis for supply chain modeling and analysis, dissertation proposal in industrial engineering, north carolina state university. . [ ] s. lertworasirikul, s. c. fang, h. l. w. nuttle, j. a. joines, fuzzy bcc model for data envelopment analysis, fuzzy optimization and decision making ( ): - . [ ] y. meada, t. entani, h. tanaka, fuzzy dea with interval efficiency, proceedings of th european congress on intelligent techniques and soft computing. eufit ' . aachen, germany, verlag mainz. ( ): - . [ ] r. mulwa, a. emrouznejad, l. muhammad, economic efficiency of smallholder maize producers in western kenya: a dea meta-frontier analysis, international journal of operational research. ( ): in press. [ ] r. mulwa, a. emrouznejad, f. m. murithi, impact of liberalization on efficiency and productivity of sugar industry: the case of mumias sugar company in kenya, journal of economic studies. in press. [ ] s. saati mohtadi, s. a. memariani, g. r. jahanshahloo, efficiency analysis and ranking of dmus with fuzzy data, fuzzy optimization and decision making. ( ) - . [ ] j. k. sengupta, a fuzzy systems approach in data envelopment analysis, computers and mathematics with applications. ( ) : - . [ ] m. zerafat angiz, a. emrouznejad, a. mustafa, a. rashidi komijan, selecting the most preferable alternatives in a group decision making problem using dea, expert systems with applications. ( ) – . [ ] m. zerafat angiz, a. emrouznejad, a. mustafa, a. fuzzy assessment of performance of a decision making units using dea: a non-radial approach, expert systems with applications. ( ) – . [ ] m. zerafat angiz, a. emrouznejad, a. mustafa, a. s. al-eraqi ( ) aggregating preference ranking with fuzzy data envelopment analysis, knowledge-based systems, in press. [ ] h. j. zimmermann fuzzy set theory and its applications, kluwer academic publishers. . pii: s - ( ) - what makes expert systems survive over years—empirical evaluation of several engineering applications jukka k. nurminen*, olli karonen, kimmo hätönen nokia research center, p.o. box , fin- nokia group, finland abstract this case study analyzes eight expert system applications that have successfully been in industrial use for a long time. we have personally been involved in the development of these applications and are thus in a good position to analyze what is important for a successful application and what kind of mistakes can be made. since the development of the applications started in – and some of them are still in use we are able to observe what has happened to those applications during their lifetime. our key observations are related to the scope of the applications, to the trade-off between usability and automation, to the role of human experts in the use and development of expert systems, on the technical solutions used, on aspects of the operation of the expert system and on the similarities between expert systems and information systems. the key findings are expressed as hypotheses for successful expert systems. the support of each application to the hypotheses is discussed. q elsevier science ltd. all rights reserved. keywords: expert systems; industrial applications; lessons learned; success factors; implementation issues . introduction in late s there was a period of high activity on expert systems, with a lot of commercial interests and expectations. as very few of the overdimensioned expectations were met the interest declined (harmon, ) but now the activity seems to be increasing again. the goal of this paper is to look back at some expert systems that were developed at that time in nokia research center, discuss their lifecycle, and see where those systems are today. this experience allows us to extract several lessons learned that are important for the success and longevity of expert systems. in this paper we would like to emphasize the aspects of practical, industrial usage of the expert system technology. the applications are real-life, deployed applications that have been in production use within nokia or its partners for years. we have personally been working on the develop- ment, deployment, and use of these applications. we feel that this is a good time for a post mortem analysis since now over a decade has passed since the applications were first deployed. it is thus possible to review what happened to those applications, what aspects were important for the future development and maintenance, and which applications or features were able to stand the test of times. application descriptions in the literature commonly discuss either prototypes or solutions that have been in use for a short time. such descriptions illustrate the possibilities of new technologies but they fail to cover issues that become visible after a longer period of serious use. grandon gill ( ) statistically studied the status of the early successful expert systems to see what had happened to those systems in a decade. only about a third of the systems that were initially reported as successes were still in use. the reasons were found to be individual and organizatorial rather than technical or economical. the important issues highlighted in that study were: coordinating es develop- ment with the it and business strategies of the firm, understanding the task to be performed by the system, recognizing the legal implication of systems, identifying user concerns and expectations, and managing developers and development responsibilities. guimaraes, yoon, and clevenson ( ) tested several hypotheses for expert system success using a sample of expert systems at e.i. dupont de memours and company, inc. most of the systems were very small and developed by one person. the study found support for seven of the nine hypothesized determinants of es success. the important factors were: problem importance, developer character- istics, shell characteristics, and with less confidence: end- user characteristics, es desirable impact on end-user jobs, and user involvement. the two factors that did not find - / /$ - see front matter q elsevier science ltd. all rights reserved. pii: s - ( ) - expert systems with applications ( ) – www.elsevier.com/locate/eswa * corresponding author. tel.: þ - - ; fax: þ - - . e-mail address: jukka.k.nurminen@nokia.com (j.k. nurminen). http://www.elsevier.com/locate/eswa support were problem difficulty and managerial support. another very similar statistical study (guimaraes, yoon, & o’neal, ) focuses on manufacturing expert systems withing ibm. duchessi and o’keefe ( ) analyzed six cases (three successes, three failures) within three companies focusing mainly on engineering applications. the success factors found in their study include management support, demon- strable business benefits, problem urgency, degree of organizational change, organizational support, and users’ personal stake in the system. millett and powell ( ) focused on the organizational aspects. they identified measures for the success of expert systems and used these measures to evaluate and benchmark the performance of one anonymous company. the critical success factors found in that study include: the type of application, importance of the system to business strategy, an early, key, successful development, existence of project champion, existence of an information technology strategy, and organizational culture. main factors contributing to failure were found to be poor project control, insufficient resources for maintenance, and financial constraints. stylianou, madey, and smith ( ) performed a field survey among expert system shell users to identify the most important features needed for expert system shells. the most critical characteristics found in that study were embeddability, rapid prototyping, backward chaining, explanation facility, ability to customize explanations, linkages to databases, and documentation compressiveness and readability. other authors have focused not only on expert systems but ai research and applications in general. hayes-roth ( ) discussed from his subjective perspective different fields of artificial intelligence, tried to find the most successful areas, and presented practical issues for real- life use. allen ( ) also uses a subjective perspective to discuss the opportunities of ai underlining the connection between ai research and real world needs. doyle and dean ( ) summarize the results of the strategic directions in computing research ai working group. among a variety of topics the paper highlights the need to integrate ai research and solutions to other research disciplines and computing systems, collaborative problem solving, and usability of the computers. the problem with all case-based research is that the results are specific to the applications and companies studied. however, the similarity of results from two companies, ibm (guimaraes et al., ) and e.i. dupont de memours and company, inc. (guimaraes et al., ), indicates that the conclusions from one organization can be applicable to others. a problem, especially with the statistical studies covering a large group of systems, (grandon gill, ; guimaraes et al., , ; parpola, ), is that the tested hypotheses are quite abstract. studies with smaller sample or narrower focus give more concrete, practical advice. measuring and collecting detailed data of application development and lifecycle without disturbing it is proble- matic. one suggested way is to retrospectively look at the data available and consider what kind of hypotheses could be tested with the data (fenton, ). this is the empirical approach we have followed here. this paper complements the previous studies by giving support to some earlier observations, refuting some other ones, and suggesting a set of new observations. the rest of this paper is structured in the following way. in section we provide a more detailed overview of the applications. section discusses our experiences and the lessons learned. section summarizes the most important observations of this study and suggests ideas for further research. . applications starting from nokia research center had a group focused on knowledge technology, which was later merged with the software technology department. the mission of the group was to develop knowledge-based applications for nokia internal use and for key customers. the emphasis was strongly on practical applications; very little basic ai research was done. table lists the most influential and widely used ones of these applications, some of which have been in use for close to years. the table also shows the domain, the lifetime, the software and hardware environment (which may have changed during the lifetime), and the current status of each application. as can be seen from table , five applications deal with planning types of activities, two with diagnostics, and one is a decision-making system. all of the applications except one (matias) work on the engineering domain. . . nicad nicad (karonen, ) was a knowledge-based system for designing plants and other complex entities. its first target application was the design of the soda recovery boilers for paper industry. a boiler plant contains thousands of components and hundreds of kilometers of pipes. the design of a new unique plant can take over person years. in nicad, the components of the plant (e.g. valves, pipes) were modeled as frames using an object-oriented approach. each frame contained attributes that described the needed features of the corresponding component. most of the attributes were geometric but there were also other types such as materials and standards. each attribute could be associated with a calculation rule that tells how the value of the attribute depends on other attributes. the local dependencies between attributes defined global dependency graphs. these graphs were j.k. nurminen et al. / expert systems with applications ( ) – two-directional and nicad behaved like a d spreadsheet. on one hand, the dependencies were used in the calculation of the value of an attribute. on the other hand, if you changed a value, the resulting modifications were propa- gated in the opposite direction in the dependency graphs. the whole plant was described as a hierarchical product structure. e.g. a power plant consists of the boiler and the building, the boiler consists of the furnace and the economizer, etc. the product structure contained also calculation rules for the numbers of components (e.g. the number of support pillars), which changed dynamically during the design based on the dependencies. the calculation rules were expressed by means of a simple language, including lisp functions and keywords like right, left, behind and above: object c is to the right of object a by units and under object b by five units. nicad was developed initially in the kee expert system environment but was later ported to c and sun. the current version of the tools, under the name designþþ, runs on windows nt platform. interfaces to odbc compliant databases and apis for c, cþþ and visual basic are supported. in a separate spin-off company, design power inc., located in silicon valley, was founded to handle the commercialization and further development of the tool. . . rft rft was a design support system for analog and especially radio frequency (rf) circuits. it was developed mainly for the needs of nokia mobile phones. rft is discussed in greater detail in ketonen, lounamaa, and nurminen ( ). rft was in a pilot usage but the complete system never reached a wide user space. one of the reasons was the lisp implementation language, which tied the system to an expensive engineering workstation. later the intelligent user interface of the system was reimplemen- ted in cþþ for the windows platform. the user interface is still used today together with the aplac simulator. rft was based on three main techniques. first, the graphical user interface allowed the direct manipulation of circuit diagrams. this was the most successful part of the system and is still in use today. secondly, a rule-based layer was used to hide the details of the tools from the users. a number of separate tools were needed for rf design at that time. since each tool had their own relatively unfriendly user interfaces the idea was to create a knowledge-based layer that would take care of transforming a user-specified problem to the appropriate table the main expert system applications developed at nokia research center in – name domain lifetime platform status planning and configuration nicad intelligent cad for engineering applications – today kee, symbolics ! c, sun, microsoft windows since continued within an independent company design power (http://www.dp.com) by the name designþþ rft electronic cad system for mobile phones – today lisp, hp ! cþþ, microsoft windows intelligent gui in use as part of the aplac simulator package. since continued within an independent company aplac solutions corporation (http://www.aplac.com) nokia planner project planning – today(?) lisp, apollo ! c, microsoft windows since continued within an independent company visolutions by the name visual planner. merged with icl cabcon cable configuration – xi plus, genesis, nicad ! c in trial use in nokia cables until dxcon telephone exchange configuration – sqlwindows, microsoft windows in production use in nokia networks until , diagnostics dmg diagnostics of radio links – lisp, apollo, sun, hyperexpert, pc in production use in nokia networks until , dxlib documentation management – today lisp, sun and cþþ, microsoft windows ! www in production use in nokia networks. dxlib model is still part of product shipment. decision-making matias loan application analysis – today(?) xi plus/guru, pc in production use at oko bank finland j.k. nurminen et al. / expert systems with applications ( ) – http://www.dp.com http://www.aplac.com form for each tool. in this way the user would be able to work with familiar schematic circuit diagrams. the symbolic manipulation at the knowledge-based layer would take care of the details, constraints, and complex syntaxes of the underlying tools. although this concept worked well on the pilot cases it was not flexible enough. a further problem was that hiding too many details was confusing to the user. the third focus area was the design synthesis. the idea was to collect design rules from experienced engineers and thus allow the distribution of the knowledge to more inexperienced persons and to geographically different locations. this turned out to be extremely difficult. creating a prototype system was relatively easy but updating and maintaining the design rules became a problem. additionally, automated design synthesis was not flexible enough since, as observed also in simon ( ), the goals and constraints are changing during the design process. the creativity of the user cannot be substituted by an automated solution. . . nokia planner nokia planner was a project management tool. the initial goal was to automate key project planning and project tracking tasks within nokia. after the initial prototype was built in lisp on symbolics the project focus shifted from automation to visualization. with a graphical user interface running on early versions of microsoft windows nokia planner provided a graphical user-interface that allowed easy manipulation of activities and their interdependencies as well as resources. an automatic feature was available for scheduling the tasks to minimize the project duration. however, it turned out that the users mostly relied on manual manipulation of the activities since this allowed them more flexibility. nokia planner was moved to an independent company visolutions that continued the development and marketing of the tool under the name visual planner. visolutions was merged with icl in . . . cabcon cabcon was a system for cable configuration (karonen, ). the design of a cable is much more complex than the simple outer sheath lets us believe. the topology (what and how many components), materials, and geometry (also d) of the cable must be determined to meet the requirements of the client including transmission capacity and quality, purpose of usage, strength, maximum allowed weight, latest time of delivery, etc. in addition to the mechanical, thermal, and electronical properties of the cable itself, also the overall production situation has an influence on the priority of alternatives. the general goal is to find out a feasible solution involving minimum costs. the objective of the cabcon tool was to increase productivity and quality of cables design and manufacturing at nokia. the subgoals were: prompt and high quality responses to customer inquires, lower costs in production, documentation of the design knowledge, standardization of delivered products, and reduced time-lag between design and delivery. the project started in , and three prototypes were implemented using three different expert system shells, respectively. exploiting the results of the complementary prototypes, a small production version was implemented in c in . the system was in limited, mainly experimental, use until . cabcon did not fully satisfy the users’ needs for two main reasons. first, the repertoire of cable structures needed is so large that the cabcon rules were not able to generate versatile enough solutions. secondly, a graphical user interface was missing which made it difficult to update, customize, and manipulate the configurations proposed by the system. these two shortcomings amplified each other: automatically generated solutions were not satisfactory and their manual improvement and fine-tuning was too clumsy. . . dxcon dxcon was a system for configuring nokia dx telephone exchanges. configuration of a dx switching system was a tedious task, and only a few experts were available. equipping required both brute force and good command of the frequently evolving design rules. dxcon was meant to be an expert tool for design, sales, and production personnel to increase the productivity and the quality of equipping and price setting of dx switching systems. some of the equipping phases were totally automated, but for the more sophisticated design tasks interactive tools were considered as the most optimal compromise. a switching system consists of rack rows, racks, subracks, plug-in units, and cables. a central concept in dxcon was an equipping model, e.g. for each rack type there was a template describing what kind of subracks can be equipped and what are the feasible locations for those subracks. for each cable we had to define its type, source and destination, routing, and length. in full-scale production use the dominant way was to reuse and incrementally modify earlier complete switching system models created by the end-users themselves or by more senior experts. dxcon was running on a standard personal computer with a relational database package called sqlwindows by gupta technologies. one of the main problems of dxcon was its speed. even after several optimization rounds some operations required about half an hour to complete. complex db schema, queries with several joins, and big db size were main reasons for the slow operation. during its lifetime starting from dxcon was in wide use and several hundreds of switching system j.k. nurminen et al. / expert systems with applications ( ) – deliveries were configured with it. its use was terminated when the importance of the fixed telephone exchange business decreased. . . dmg fault diagnosis of complex technical equipment is always a big challenge. one solution for the problem has been model-based expert systems, where a predefined decision system guides the analysis and is able to identify the fault. dmg (tyrväinen, ) was used to create and edit decision trees that were used in the core of the diagnosis system. the dmg system had two modules: an engineering workstation, where the equipment model was created and where a decision tree was extracted from the model. the tree was then transferred to the end user environment, which was a standard pc with an expert system shell that was able to execute the decision tree. the system was used at radio link manufacturing. the link equipment was tested and fixed before they left the factory. . . dxlib dxlib (tyrväinen, saarinen, & hätönen, ) was a system for document management and information retrieval from documentation of dx based network elements. its main components were a text analyzer that produced sgml tagged formatted documents and a browser that linked documentation to the logical model of the element. the objective of the system was to assist human expert in finding relevant information. like dmg the system consisted of two modules. an engineering workstation, where a logical model was created, was used to link the model and relevant texts together. the model and the documentation were then transferred to a pc environment, where they could be browsed. the system has been used ever since its first trial in . the model is an optional feature of the electronic network element documentation package that accompanies network deliveries. . . matias matias (kontio, ) was an expert system for loan application analysis that was developed for oko bank, which at that time was an important customer of nokia data. it contained knowledge about farming loans in finland. the regulations for such loans including state subsidies were quite complex and matias was meant to speed up the loan decision process. containing a few thousand if-then rules divided into multiple modules matias was a large expert system at its time. most of the knowledge used by matias existed already on documents about the loan policies and regu- lations. however, a large number of rules was needed to codify this knowledge. the large size of the rulebase resulted into the usual problems of verification and validation of the rulebase as well as difficulties in maintaining the rulebase. . characteristics of successful systems we have collected our main observations in table . the ‘hypothesis’ column contains properties or statements that are characteristic of successful expert systems. these hypotheses are either derived from literature or are based on our own observations. the ‘source’ column provides references to literature that has suggested a given hypothesis. it is difficult to track the ultimate origin of each observation or hypothesis in the literature. therefore we have mentioned those studies where we have encountered a mentioning of a given hypothesis. also different authors have formulated their findings in different ways. we have tried to understand what has been the spirit of the observation when listing which authors have discussed each hypothesis. the ‘evidence of our case study’ column indicates general level of support this study gives for each hypothesis. the following columns contain more detailed analysis for each application. a ‘ þ ’ means that the application supports the hypothesis, a ‘ ‘ means it refutes the hypothesis. an empty value means that the aspect was not relevant or observed for the application or that the conclusion is not clear. the observations are divided into four categories: domain, development, operation, and general. the hypoth- eses and the case study evidence for each of them are discussed in greater detail in the following sections. . . domain . . . narrow scope hayes-roth ( ) suggests that narrowing the scope is important for expert system success. naturally it is easier to develop a system with a narrow scope. however, the impact of such a system tends to be low because the solution is too limited. fig. shows the difference between narrow and wide scope applications. with a narrow scope it is possible to develop a system that is able to generate high quality solutions. these solutions can often do things better than average experts. however, it seems that the law of diminishing returns applies: the higher the solution quality is, the bigger marginal effort is needed to increase it. this means that pushing the solution quality above the acceptable level requires much more development than settling with a more modest level. the drawback of narrow scope is that it is possible to cover only a small part of the problem domain with such systems. although a large enough number of narrow j.k. nurminen et al. / expert systems with applications ( ) – table case study evidence for the hypotheses hypotheses source evidence of our case study nicad rft cabcon dxcon dxlib dmg matias planner domain narrow scope hayes-roth ( ) mixed þ þ focused objective hayes-roth ( ) no stability of environment hayes-roth ( ) no high degree of automation hayes-roth ( ) mixed þ þ þ high degree of repetition hayes-roth ( ) yes þ þ þ þ þ þ þ þ small project hayes-roth ( ) yes þ þ þ þ þ þ þ þ development users prefer usability over automation nurminen, karonen and hätönen ( ) yes þ þ þ þ þ expert systems complement rather than replace human experts millett and powell ( ) yes þ þ þ þ þ þ þ early benefits to experts themselves are important nurminen, karonen and hätönen ( ) yes þ þ þ þ þ þ ai should be an embedded part of a bigger system duchessi and o’keefe ( ) and millett and powell ( ) yes þ þ þ þ þ þ þ simple, straightforward solutions work best nurminen, karonen and hätönen ( ) yes þ þ þ þ þ þ þ if-then rules considered harmful nurminen, karonen and hätönen ( ) yes þ þ þ lots of custom work hayes-roth ( ) yes þ þ þ þ þ þ þ fast and agile development is important duchessi and o’keefe ( ) yes þ þ þ þ þ þ þ knowledge-based applications are based on domain specific application engines and generators nurminen, karonen and hätönen ( ) yes þ þ þ þ monolithic applications are sometimes better than knowledge-based applications nurminen, karonen and hätönen ( ) yes þ þ þ rules of normal sw development apply duchessi and o’keefe ( ) and millett and powell ( ) yes þ þ þ þ þ þ þ operation move towards mainstream software and hardware platforms nurminen, karonen and hätönen ( ) yes þ þ þ þ þ þ successful systems tend to move out from the company nurminen, karonen and hätönen ( ) some þ þ þ general the difference between expert systems and information systems is small duchessi and o’keefe ( ) yes þ þ þ þ þ þ þ j .k . n u rm in e n e t a l. / e x p e rt s y ste m s w ith a p p lic a tio n s ( ) – systems could in theory cover the whole problem domain, in practice there are not enough development resources. when the scope is wider the level of expertise is typically smaller. in that case the expert system is not targeting complete solutions. instead, it relies on humans to supplement the draft or partial solutions provided by the system. in this way it typically is easier to cover a much larger area of the problem domain. a comparison of the two graphs in fig. shows that total work for human experts is less when the expert system covers a wider problem area with rough level of detail. at the same time the average quality of solutions tends to be higher. the consequence is that the system cannot work autonomously and a human expert is continuously needed. most of the applications of this study had a wide scope. only the loan decision-making tool (matias) and the radio link diagnostics tool (dmg) had a narrow scope. when the problem domain is large, like for the planning applications, the wide scope allows some support for most of the tasks. . . . focused objective hayes-roth ( ) states that focusing on a single objective, such as to reduce time, to reduce cost, or to produce a better design, is beneficial. our experience strongly contradicts this. in all of our applications we have had multiple goals. typically we have tried to improve quality, increase speed and save time at the same time. the need to emphasize multiple goals is especially evident in planning tasks. simon ( ) observes that design work has multiple goals and constraints and that their priority and importance changes during the design process. our experience is very much in line with this. for any reasonable complex planning task it is impossible to list all of the objectives in advance. focusing the design on the optimization of a single criterion tends to create plans that are not feasible in practice. . . . stability of environment stability of environment is another issue where our experience strongly contradicts (hayes-roth, ). a stable environment helps the implementation but the assumption of a stable environment is too idealistic. the shortening release cycles of all kinds of products make the environment even more dynamic than what is has been in the past. it is essential that the expert systems are useful in changing environments. otherwise they would either be outdated or their maintenance would be a constant problem. systems with a narrow scope have a high risk in this respect. as can be seen in fig. if the problem domain moves slightly (e.g. with the introduction of a new technology) it can happen that the narrow scope system falls out from the problem domain. for the wide scope systems this risk is much smaller. . . . high degree of automation high degree of automation may seem like a desirable feature. as described in hayes-roth ( ) this would cause savings in the operation of the system. in this issue our evidence is mixed. the narrow scope systems, matias and dmg, provide a high degree of automation. in dxcon and nicad some parts are highly automated while others require a lot of user involvement. after the planner had created the configuration dxcon took care of hardware details and thus saved work and reduced errors. nicad automated a lot of the calculation even if human experts made the design decisions. in the other applications the general level of automation was not very high. we can see several explanations for these observations. first, matias and dmg were intended for non-expert users. all the other applications aimed at improving the performance of an expert user, which seems to be common in engineering applications. secondly, increasing the degree of automation requires increasingly more development effort. when the problem fig. . narrow vs. wide scope. j.k. nurminen et al. / expert systems with applications ( ) – complexity and width increases reaching a high degree of automation may become infeasible goal. in terms of fig. the question is how much width is sensible to sacrifice to increase automation. one practical way is to develop some modules with high degree of automation to automate routine tasks and rely on the end-user in the more difficult tasks. modules containing calculation or detailed but straightforward computation are good candidates. . . . high degree of repetition high level of repetition is clearly useful. it allows the benefits from the system to be drawn continuously and thus makes the investment in the system worthwhile. in addition to the benefits that are gained from using the system the high degree of repetition is also useful for the development. more repetition means more feedback to the development, more need for new features, and more incentive to fund the further development and maintenance. . . . small project our expert system projects have been small with less than , typically – , full-time developers plus part-time experts. we do not have hands-on experience with the larger expert system projects. with our data we can only conclude that small projects work but the data does not say anything about large project. our subjective feeling, based on experience with expert system and large-scale software development projects, is that large-scale expert system projects should not be more difficult than any large software projects as long as the expert system parts are developed as compact, well- defined, and reasonably independent components. . . development . . . users prefer usability over automation finding the proper balance between usability and automation seems to be an issue in most projects. since the applications are developed with limited resources it is important to focus the development to the most useful direction. the experience with matias and with some modules of dxcon indicates that when the degree of automation is high the usability is not considered very important. if the automatically generated solutions are fully correct the user has no need to analyze and interpret the results. the situation is different in applications where full automation is not feasible. in that case usability is a major issue since users need to control the application, understand the solutions, analyze the trade-offs, and manually adjust the solutions until they satisfy the needs. this seems to be the most common case. automation is valued but only up to a certain level. the users want ultimately to be in control and understand the solution before they can accept it. this division also mirrors the difference between expert and ordinary users. matias brought the loan knowledge to the customer agents in bank offices. dxcon had two classes of user: senior experts who defined the component configurations and less experienced expert users who utilized the expertise. in the other applications the users were typically the experts themselves. it seems that in tools that are intended for the experts the usability is very important. the expert can compensate for missing auto- mation by using his experience and competence. the ordinary user is more or less blindly relying on the results of the tool and therefore is not very interested, or competent, to analyze the rational behind the results. graphical user-interfaces naturally have a major impact on usability. in fact, the graphical user interfaces seem to be one of the most important results of long-term practical value of the ai boom in late s. today the graphical user interfaces are commonplace but years ago the expert systems were among the first applications with intuitive user-friendly graphical interfaces. . . . expert systems complement rather than replace human experts a key characteristic of our applications has been that they are intended for expert users such as product designers or electrical engineers. in such a role the tools support, rather than replace, the expert. according to our experience for any non-trivial planning task it is not possible to create completely automatic solutions. an expert user is essential to handle changing and unexpected needs, multiple objectives, and other similar aspects where the human flexibility and a wide under- standing are important. systems which combine the strengths of both humans and computers, intelligence amplification systems (brooks, ), seem thus to be the most promising direction for practically successful applications. millett and powell ( ) made a similar observation that ‘one reason for success is that expert system has not attempted to undertake the whole task, merely a part of it that is time consuming for staff. this frees experts for more complex tasks not handled by the system’. requiring the user involvement may seem like a failure from a puritanist ai perspective but from the practical perspective it solves many problems and results into less complex systems. requirements for the completeness and correctness of the system can be less strict. the system can consist of a set of basic tools that are controlled by the user. changes and new demands do not necessarily require a modification to the system but can be handled by the end user. as a result the application development is easier and the applications are more adaptive to external changes. . . . early benefits to experts themselves are important the experts with the relevant knowledge and under- standing of the problem are key to expert system success. typically the experts are busy, which is one reason why it makes sense to develop the expert system. there is often j.k. nurminen et al. / expert systems with applications ( ) – a conflict between the short-term and long-term priorities when the expert schedules his time between different tasks. our experience suggests that one very useful practice to involve the expert is to develop the system in such a way that it offers benefits almost immediately to the experts themselves. it is not enough to satisfy the end-users. it is equally, and sometimes more, important to satisfy the expert to be able to create a good enough system to be delivered to other experts and end-users. when the experts can directly influence the development their attitudes tend to be more positive towards the new systems. early participation in development reduces the often-perceived risk that the expert will be replaced by the expert system. instead active involvement emphasizes the importance of the expert and the tool makes the expert more competent. duchessi and o’keefe ( ) state a more general observation that a system that directly benefits a user fosters operational use. while we agree with their observation we feel that emphasizing early benefits and to the experts themselves is important. otherwise the development easily slows down and stops. at least one of the successful cases that (duchessi & o’keefe, ) analyzed seems to confirm this. . . . ai should be embedded part of a bigger system all of the applications, except matias, show that expert system modules cannot be handled in isolation. to be useful they have to be embedded to larger software systems. the expert system modules perform key tasks and provide the intelligence of the application but without the other parts the intelligent parts would not be useful alone. millett and powell ( ) found that stand-alone systems made users unhappy by requiring existing data to be re-keyed. two of the three successful expert systems in duchessi and o’keefe ( ) were actually database applications where the expert system module was ‘icing on the cake’. the first implication of this is that expert system modules are most useful when implemented in the same environment and same tools as the rest of the application. often this means the use of a standard programming language like cþþ or database centric solutions, like in dxcon. it still makes sense to specify the expertise in special format in definition files that are either loaded to the application or compiled and linked as part of the application. another implication is that in the development a lot of effort has to be spent on the non-ai parts. this work involves, for instance, the graphical user interface as well as interfaces to external systems, and can easily exceed the effort needed for the expert system parts. in the project planning it is thus important to allocate enough effort to all parts of the system and avoid the mistake of focusing only on the key expert system parts. . . . simple, straightforward solutions work best if the ambition level is too high it easily happens that the system uses sophisticated techniques and solves limited subsets of the target problem very well, but it is not able to handle the real size problems. in almost all of the applications in this study the experience was similar. the objectives for the degree of automation were at the start much higher than what was accomplished in the end. the advanced ai modules were pruned and the mainstream technology modules started to dominate the development. better results can be achieved when the system focuses on solving the practical problem. millett and powell ( ) found that successful organizations were less concerned with pushing the technology and more inclined to concentrate on the application. a typical sin on the expert system area has been to try to force a certain solution to a problem. instead it is important to consider which solution approach would the optimal one for a particular problem. although expert systems are good solutions for some problems, for other ones e.g. the use of optimization and other mathematical algorithms could be better choices. developers with a wide understanding of possible solutions techniques have a key role. naturally finding developers with such broad experience is difficult. . . . if-then rules considered harmful the analysis of the applications of this study indicates that if-then rules are not the right way to represent engineering knowledge. this finding is in contrast with the study (stylianou et al., ) that found support for backward chaining to be the third most important feature in expert system shell selection. matias, the only non-technical expert system of this study, was the only deployed application that successfully used if-then rules. a few of the other applications also experimented with using if-then rules on some parts but these trials did not give good results. technical knowledge deals with artifacts and their relationships, formulas, and numeric values that are more easily expressed as classes and objects. rules are more suitable for describing knowledge on the area of legal reasoning, such as the terms of loan approval. another observation is that the maintenance of if-then rulebases is more difficult than for most other knowledge representations. instead of the general if-then format a special application specific structure is useful for easier development and maintenance. the preexisting solutions available in application development tools are seldom suitable as such. finding the right structure is very similar to software architecture design, which is a key step for successful software development. . . . lots of custom work the large individual differences between developers (see e.g. mcconnell ( )) are likely to be even bigger in expert system development than in normal software development. the reason for this is that, being in the intersection of ai and j.k. nurminen et al. / expert systems with applications ( ) – information technology, expert system development requires more creative solutions than normal software development. the task does not only require good software development competence but also good understanding about the applicability, limitations, and usefulness of different solutions techniques. because many expert systems are pushing the state of the art it is not possible to blindly follow the solutions used in other systems. . . . fast and agile development is important in the engineering domain the expertise is developing on a fast speed. this has two implications. first, rapid development is important to bring the solution into use at the right time. if the development time is too long then the system may not be up to date anymore when it is ready. secondly, without proper maintenance and adequate tools for it the system very soon becomes obsolete. fast and agile development process is thus important. at early development phases flexibility is needed to discover the way to approach the problem and to find the right architectural solution. also, as discussed above, it is important to bring the tool early into use to ensure the expert commitment and feedback and continue the devel- opment in an iterative and incremental fashion. although duchessi and o’keefe ( ) do not explicitly state the importance of this it seems that a user driven phased development has been used at least in the successful systems that they analyzed. the agile development is another area besides graphical user interfaces where the expert system community has done pioneering work. the incremental, customer driven style of expert system development has only recently become accepted, and increasingly popular, in software community at-large (see e.g. cusumano and yoffie ( )). . . . knowledge-based applications are based on domain specific application engines and generators as shown in fig. the applications of this study can be divided into three groups † monolithic applications have no clear difference between knowledge and application logic. monolithic applications are very close to general is applications. † knowledge-based applications separate the knowledge and the processing of knowledge (engine). this is the traditional model for expert system applications. the same engine can be used with different knowledge to create different applications. maintenance of the appli- cation is supposed to be easy since the changes are mainly made to the knowledge part. † application generators are not intended for end-users but for developers. their main purpose is to assist in the application development. since the development and maintenance of a knowledge-base is typically difficult, the generator applications are often essential in making the knowledge bases more robust, easier to understand, and faster to develop. the generators typically work by representing the knowledge to the developer and to the expert in a form that is easier to understand and manipulate. in the application the primary goal of knowledge representation is efficient execution, in the generator it is human comprehension. this study suggests that the textbook solution to use a general-purpose expert system shell and to concentrate on the knowledge base in the application development does not work. there is a need to create very specialized, domain specific tools for the engine (nicad) or for knowledge base development and maintenance (dmg, dxlib). such tools take advantage of the special structure of the knowledge in a domain and make the knowledge more manageable by constraining the knowledge to a more compact, consistent and modular format. matias was the only application developed with an expert system shell but the experience with it also triggered the development of knowledge engineering tools (parpola, ) to better handle the knowledge acquisition process. dxcon was the only application that did not support this observation. an explanation for this is that dxcon was a database centric application and stored the knowledge in fig. . applications and their types. j.k. nurminen et al. / expert systems with applications ( ) – the form of configurations components to relational database tables. this kind of open architecture allowed easy addition of new modules developed with different technologies. own development effort is needed to find the right structure for the knowledge but after that initial effort the rest of the development becomes a lot easier. developing the engines and generators requires major work. the knowledge-based system creation is thus only partly about the knowledge acquisition. it is also about software development to create suitable tools that can be efficiently used to build the applications. for instance, at the beginning most of the work on the nicad application had to be spent on developing the engine. the situation may have changed and today there are more tools that can be used as building blocks for the applications. rather than expert system tools such software is marketed as domain specific automation tools. . . . monolithic applications are sometimes better than knowledge-based applications for most end-users and applications the difference between monolithic applications and knowledge-based applications is small. a major difference can arise if the normal operation of the system requires that experts dynamically update the rules. the division has a bigger effect for the development and maintenance of the applications. knowledge-based appli- cations are often seen as the orthodox way to develop expert systems. also in our case that was the starting point of all the applications. however, as the development progressed planner, rft, and cabcon shifted to the monolithic style. this was mainly caused by the fact that the importance of application specific knowledge turned out to be less important than initially expected. knowledge-based application have their price. the development of a knowledge-based application, in our case, required also the development of the engine and generator tools. reaching the needed generality with such tools requires development effort. separating the knowledge from the engine also tends to reduce the execution speed. our observation is that the selection between monolithic and knowledge-based application depends on the type of knowledge. if the knowledge is quite static and not growing, if the amount of knowledge is not very large, and if its representation with mainstream information technology is easy then a monolithic application can be a good choice. proper software architecture makes it also possible to modify the knowledge. naturally it is harder than in the case of a knowledge-based application. further characteristics of successful monolithic appli- cation seem to be a wide and general application domain, high importance, and expert users. . . . rules of normal sw development apply our experience suggests that there is very little difference between software development and expert system develop- ment. most of the successful solutions in our applications are very close to normal software development. likewise the process of developing an expert system and developing normal software is very similar. there are differences in emphasizing the importance of different steps rather than the actual work. our experiences confirm those of duchessi and o’keefe ( ) who stated that factors leading to success in information system and decision support system develop- ment are also important for expert system implementation. the normal software process activities like development, testing, and maintenance easily become very difficult when a larger knowledge base is needed. this is especially problematic if the modularity of the knowledge is low. if the knowledge is stored in rulebases with little internal structure the number of alternative execution paths becomes very high. it is no longer possible to understand how the rules interact, is the rulebase complete, or what happens when something unexpected happens. testing all of the execution paths is clearly beyond practical limits. moreover, the maintenance work easily is very difficult. maintenance is the biggest consumer of software develop- ment resources. some estimates assume that % of software lifetime cost is spent on maintenance (boehm & basili, ). millett and powell ( ) observe that expert systems are no different from conventional systems in that they require maintenance. to us it seems that the maintenance of knowledge bases is even more expensive. first, the expertise tends to change frequently. secondly, the complexity of expertise and its presentation in knowledge bases makes incremental modifications difficult. finally, low modularity may result into the need to retest the whole application after each small modification. it can also happen that different experts view the functionality from different angles. it is thus possible that a rulebase that is developed with one expert and is later maintained by another is no longer consistent. one of the big successes in software engineering has been object-oriented programming and the software components. they have made the software reuse a practical issue. how is the reuse of expertise? for successful reuse the expertise would need to be packaged into some kind of component. moreover, a mechanism would be needed to control, expand, and modify the behavior of the components from the outside. at least rulebases are quite sensitive to small changes. already a change in a single parameter value in one rule requires a thorough understanding about the possible effects of the change. other domain specific formats are potentially better but still there are major difficulties for black-box reuse. domain specific engines and generators are more applicable for reuse than the knowledge itself. . . operation . . . move towards mainstream software and hardware platforms the hardware platform tended to shift from specialized computers, symbolics, to unix workstations and personal j.k. nurminen et al. / expert systems with applications ( ) – computers running microsoft windows. the applications that were initially developed with lisp or expert system shells were converted to c or cþþ. there were several reasons for this shift. for widely used applications the software and hardware platform cost is essential. furthermore, with the rise of the personal computer, users tended to prefer to do all of their work on single computer using tools that had a familiar look-and- feel. the mainstream environment also allowed the creation of open interfaces to link the expert system applications with other applications. it also turned out that the platform support for expert system tools was not adequate. the companies providing specialized ai platforms, lisp programming environments, and application generators were working on a niche market. the volume of the market was not large enough to guarantee the high level of robustness, development speed, and support that is needed for serious products. moreover, most of the expert system environments were closed systems and did not easily allow the use of normal, mainstream development tools such as version control systems, makefiles, debuggers, etc. even if porting the application to new environments caused extra work this approach was, in fact, very productive. the developer-oriented environments and application generators allowed a fast start and enabled early end-user involvement. when the user needs were understood and the appropriate solution found in this way, it was relatively straightforward to reimplement the relevant parts of the solution in another environment. another benefit of the platform change was that it was a good opportunity to improve the speed and robustness of the application. when the needs were clear it was possible to choose architectural solutions that allowed efficient and future-proof applications to be written. . . . successful systems tend to move out from the company our data gives some support to the observation that successful systems often lead into a spin-off company that focuses specially on the further development and mainten- ance of the expert system or the platform tools. even if the balance is even, three cases supporting and three refuting the claim, the applications supporting the claim are the most successful ones in terms of user base and longevity. there can be a result of numerous factors in the operating environment and in strategy of our company that explain this observation. expert systems, in their role as support tools, improve the operational efficiency but are not the primary products of a larger company. their benefits for the main business are indirect and hard to quantify. therefore, with increasing pressure to focus on the core business, the development and maintenance of such system may not satisfy the profit requirements. application development requires that a lot of effort is spent on monolithic applications or on domain specific engines and generators. it thus becomes attractive to reuse the platform in order to divide the platform development cost between several applications. such a generalization, which e.g. happened in nicad, results into the creation of application specific environments that can be used as a basis of a product family. in our case the monolithic application covered a wide domain (project planning) and important special area (analog circuit design) and therefore had a large market and good business potential. a dedicated spin-off company may be in a position to handle the further development and marketing of such tools. while the mother company is mainly interested in the applications, the spin-off company can focus on the expert system technology, develop a variety of solutions around the platform, and market them to different customers. it also seems that the platform technology is more stable and easier to maintain than the actual applications. . . general . . . the difference between expert systems and information systems is small duchessi and o’keefe ( ) expected that many of the factors that lead to successful implementation of other systems are also important in expert system implementation. according to our experience many of the organizational and system development issues are common between expert systems and information systems. first, it is important to understand that the applications are essential, not the technology. secondly, it is important to develop the applications to solve the right problems at the right time. millett and powell ( ) state that it is important that an expert system project tackles a major business problem at an early stage. duchessi and o’keefe ( ) found that problem urgency and growth of the underlying application area are important factors for expert system success. thirdly, the management support, matching the solution to the organizations needs, and right organization support are important (duchessi & o’keefe, ). our experiences support all of these observations. for the expert system development it would be important to utilize all the accumulated experience of information system development. actually, we feel that one of the mistakes of expert system development may have been to view it as a distinctive area and develop own processes, tools, and methods independently from the information systems. potentially, better results would have been achieved if the problems had been approached from the information system perspective and the special expert system techniques would have been utilized only when appropriate. j.k. nurminen et al. / expert systems with applications ( ) – . conclusions the domain of engineering applications is complex: technology is changing, tasks are open-ended, and multiple, often implicit, goals are common. traditional, standalone expert systems have problems to cope with these challenges. contradicting previous studies, this study suggests that wide scope, multiple objectives, unstable environment, and moderate degree of automation are important considerations for expert system development. the user of an engineering expert system typically is and has to be an expert by him/herself. expert systems should complement rather than replace human experts. particular attention should be given to usability, which expert users often consider more important than automation. part of good usability is that the expert systems are embedded and integrated parts of larger information systems. implementation of such systems has a lot in common with normal software development. in fact, most of the systems were ported to mainstream software and hardware platforms during their lifecycle. specially developed domain specific application engines and generators turned out to be highly useful. general-purpose knowledge representations, in particular if-then rules, did not suit well for engineering applications. fast and agile development was found to be important to cope with the changing environment. experts, both as developers and users of the systems, are in a crucial role. organizing the development in such a way that early versions of the systems gave benefits to the experts themselves, was a key means of getting the experts involved deeply enough in spite of their overload with other urgent business-critical tasks. the problem with empirical evaluation is that a single study provides only a piece of evidence. the analysis of multiple systems and the long time period increase the confidence of the observations of this study. however, numerous factors, like the culture of our development group, may bias the results. moreover, it is not clear how sensitive the results are to the type of industry, application domain, or point-of-time. the findings of this study are based on observations about eight successful applications. it would be interesting to deepen the analysis and look for more fundamental reasons behind these observations. the development of such theory would give more confidence to the observations and provide a general framework where the results of case studies could be positioned. references allen, j. f. ( ). ai growing up: the changes and opportunities. ai magazine, winter, – . boehm, b., & basili, v. r. ( ). software defect reduction top list. ieee computer, (january), – . brooks, f. p. ( ). acm allen newell award acceptance speech. communications of the acm, (march). cusumano, m. a., & yoffie, d. b. ( ). software development on internet time. ieee computer, (october), – . doyle, j., & dean, t. ( ). strategic directions for artificial intelligence. acm computing surveys, ( ), – . duchessi, p., & o’keefe, r. m. ( ). understanding expert systems success and failure. expert systems with applications, ( ), – . fenton, n. ( ). conducting and presenting empirical software engineering. empirical software engineering, , – . grandon gill, t. ( ). early expert systems: where are they now? mis quarterly, ( ), – . guimaraes, t., yoon, y., & clevenson, a. ( ). factors important to expert systems success—a field test. information and management, ( ), – . guimaraes, t., yoon, y., & o’neal, q. ( ). success factors for manufacturing expert system development. computers and industrial engineering, ( ), – . harmon, p. ( ). object-oriented ai: a commercial perspective. communication of the acm, ( ), – . hayes-roth, f. ( ). artificial intelligence—what works and what does not? ai magazine, (summer), – . karonen, o. ( ). intelligent cad. norwegian artificial intelligence society (nais) seminar. karonen, o. ( ). experiences with different application generators in cable configuration (in finnish, original title projekti jossa jouduttiin vaihtamaan kehitintä). blanko’ conference. ketonen, t., lounamaa, p., & nurminen, j. k. ( ). an electronic design cad system combining knowledge-based techniques with optimization and simulation. in j. s. gero (ed.), artificial intelligence in engineering: design. amsterdam: elsevier. kontio, j. ( ). matias: development and maintenance of a large but well-defined application. expert systems with applications, ( ), – . mcconnell, s. ( ). code complete. redmond: microsoft press. millett, d., & powell, p. ( ). critical success factors in expert system development: a case study. proceedings of the conference on acm sigcpr/sigmis conference, – . parpola, p ( ). object-oriented knowledge acquisition. licentiate thesis, university of helsinki. simon, h. a. ( ). problem forming, problem finding, and problem solving in design. in a. collen, & w. w. gasparski (eds.), design and systems: general applications of methodology (pp. – ). new brunswick: transaction publishers. stylianou, a. c., madey, g. r., & smith, r. d. ( ). selection criteria for expert system shells. communications of the acm, ( ), – . tyrväinen, p. ( ). dmg—object-oriented iterative knowledge acqui- sition for the generation of diagnostic hypertext systems. lisp pointers, ( ). tyrväinen, p., saarinen, p., & hätönen, k. ( ). domain modelling for technical documentation retrieval. in h. kangassalo, h. jaakkola, k. hori, & t. kitahushi (eds.), information modelling and knowledge bases iv (pp. – ). ios press. j.k. nurminen et al. / expert systems with applications ( ) – what makes expert systems survive over years-empirical evaluation of several engineering applications introduction applications nicad rft nokia planner cabcon dxcon dmg dxlib matias characteristics of successful systems domain development operation general conclusions references copyright: © elsevier science. reprinted with permission from expert systems with applications , no. , pages - . pii: - x( ) - a reduction methodology for a differential diagnosis expert system akram salah d e p a r t m e n t o f c o m p u t e r s c i e n c e , c a i r o u n i v e r s i t y kevin d. reilly d e p a r t m e n t s o f c o m p u t e r a n d i n f o r m a t i o n s c i e n c e s a n d b i o s t a t i s t i c s a n d b i o m a t h e m a t i c s , u n i v e r s i t y o f a l a b a m a a t b i r m i n g h a m a b s t r a c t the simple production rule representation is generalized by adding p r o g r a m s to a management s y s t e m that manipulate rules in a rule-based system. b y adapting this methodology, a single generalized rule can represent a group o f simple ones. then programs are e m p l o y e d to satisfy the general rule in a partial way while recursively reducing a decision p r o b l e m into smaller ones o f the s a m e nature until a decision is made. it is s h o w n that the reduction m e t h o d is m o r e efficient than the simple rule approach and that it m i n i m i z e s the n u m b e r o f rules used to express a problem. the concept o f using a m a n a g e m e n t program to manipulate a set o f rules is emphasized through solving a p r o b l e m in a differential diagnosis expert system. a comparison between the n u m b e r o f rules e m p l o y e d to express a p r o b l e m is made to s h o w advantages o f the reduction m e t h o d o l o g y over the simple rule representation. k e y w o r d s : production systems, expert systems, reduction algorithm, prolog, decision tables, relational databases, medical diagnosis i n t r o d u c t i o n much recent research focuses on computer systems that facilitate methodolo- gies simulating experts' knowledge-based decision-making strategies (see, for example, [ - ]). the most popular current way to create such systems is to incorporate large amounts of "domain-dependent" knowledge acquired from experts. because experts often express their decision-making processes in sets of address correspondence to kevin d. reilly, departments ~ o f computer and information sciences and biostatistics and biomathematics, university o f alabama at birmingham, birmingham, alabama . international journal of approximate reasoning ; : - © elsevier science publishing c o . , inc. vanderbilt a v e . , n e w y o r k , n y - x / / $ . core metadata, citation and similar papers at core.ac.uk provided by elsevier - publisher connector https://core.ac.uk/display/ ?utm_source=pdf&utm_medium=banner&utm_campaign=pdf-decoration-v akram salah and kevin d. reilly if-then rules, rule bases are devised ( h a y e s - r o t h [ ]). such systems are usually called knowledge-intensive rule-based systems. a knowledge-intensive rule-based system consists mainly o f a set o f rules describing different decision situations in the p r o b l e m under question, together with actions to be taken in each case. a rule in such a system is represented as a structure typically in the f o r m a l & a & • • • & a n - ' c such a rule is interpreted as follows: if ax and a and . . . and a n are true simultaneously, then consequently c is true. the left side o f a rule contains a conjunction o f atoms called c o n d i t i o n s and the right side is called a c o n c l u s i o n o r a c o n s e q u e n c e (salah and y a n g [ ]). i f an expert expresses his o r her decision-making process as a collection o f simple if-tben rules, each o f them can be represented directly as stated above. a p r o b l e m arises if an expert expresses rules in a less explicit w a y o r in s o m e f o r m such as a function o v e r a set o f rules. t h e n it is the responsibility o f the rule acquirer, w h e t h e r p e r s o n o r machine, to decide upon a representation o r to p r o v i d e s o m e kind o f control on processing such rules. in this article we s h o w an a p p r o a c h that can facilitate a generalization o f simple rule representation. this article is part o f a larger study that has b o r r o w e d concepts f r o m relational database systems (rdbss), such that rules are stored in a rule base and then a m a n a g e m e n t s y s t e m retrieves the rule u n d e r question. t h e system exploits a n u m b e r o f features studied previously (reilly et al. [ ], yang [ ], reilly et al. [ ]), w h e r e key c o n c e r n s have been incorporation into a p r o l o g f r a m e w o r k (kowalski [ ], c l o c k s i n and mellish [ ]) o f k n o w l e d g e representa- tions and r d b s s ( b r u y n o o g h e [ ]). it is shown that the a p p r o a c h increases the efficiency o f a rule-based system. the problem t h e p r o b l e m arises in a differential diagnosis expert s y s t e m where conditions are either s y m p t o m s , observations, o r test results gathered by a physician to be used to derive a conclusion, which in this case is a disease o r a class o f diseases. rules used to derive such a conclusion would typically be in the f o r m o l & o & . • . & o n - ' d where each o i f o r _< i _< n is an observation, and d is a disease o r class o f diseases. ( f r o m here on, we refer to any a t o m on the left side o f a rule as an observation (an observation can be a test result o r a s y m p t o m ) and on the right side as a disease. thus, this rule is read as follows: if all observations t h r o u g h n exist simultaneously in a patient, then this case can be diagnosed as d . ) o u r p r o b l e m arises when a g r o u p o f rules is expressed within a single if-then reduction algorithm in expert diagnosis statement. our particular concern is this: given a set o f n observations and a conclusion d such that if any k-subset o f observations o f n holds, k _< n, then d can be concluded. that is, any k observations o f {o & & . . . & o , } - - , d a simple example o f such a case is the c o m m o n cold, where there are about observations and any o f them (existing simultaneously) establish the diagnosis. examples o f a similar nature occur in rheumatic diseases (more discussion is provided below). actually, this is a generalization o f a rule application. the special case in which k = n defines the " n o r m a l " rule structure, that is, the case in which all the conditions have to be satisfied to derive the consequence. a more formal view o f this problem is as follows. in a rule system there is a set o f conditions for each decision situation, each condition having a domain o f values. the left side o f any rule represents an element in the cartesian product o f the domains o f these conditions. the case here m a y be conceptualized as having one condition with one domain o f observations, say, o with length n, such that i f any k-subset o f o with k < n occurs simultaneously, then the diagnosis is established. this expresses a set o f rules, each one having a condition part c e o k, where k < n, and the same consequence d , which is the disease under consideration. t o represent this situation within an expert system, we examine two alternatives to set the stage for subsequent comparison. . the single if-then statement is re-expressed as a set o f simple rules. each such simple r u l e c o n t a i n s k observations on its left side and d on its right side. needless to say, the resulting number o f rules consumes much m e m o r y space, complicates the search when the system is applied, and reduces the efficiency o f the system. . the production system is extended such that if the " a n y k out o f n " formulation is expressed, it can be handled automatically. note that this problem differs f r o m those representations o f " i n e x a c t r e a s o n i n g " or uncertainty (prade [ ], rosenbloom et al. [ ]) in which subsets o f conditions are used to derive a c o n s e q u e n c e - - f o r instance, probabilistically, fuzzily, or using weighting schemes. r e d u c t i o n m e t h o d the reduction method is based on viewing rule-base systems as a set o f rules together with p r o g r a m s that manage such rules. this view enables us to add program code to the management system such that it can extend the simple representation o f production rules. here, we apply this methodology to enable a direct representation o f the generalized f o n n discussed above. we denote a p r o b l e m as " a k / n diagnostic p r o b l e m " when the diagnosis is akram salah and kevin d. reilly dependent on n observations such that if any k o f them are found to exist in a case, then the diagnosis is established. f o r example, the case discussed by weiss and kulikowski ([ ], p. ff.) in diagnosis o f rheumatic diseases such as mixed connective tissue disease ( m c t d ) involves observations. i f any o f these exist in a patient, then he or she definitely has a rheumatic disease. using our terminology, we say that this is a / diagnostic problem. a reduction algorithm t o solve a k/n diagnostic problem using the reduction methodology, we p e r f o r m the following: . pick any k symptoms. . name them temporarily t . . . . . tk. . check decision table (k) with results for the k symptoms. . the output o f the decision table is ts. . the problem now is ts/r, where r = n - k. . for any /r diagnostic problem: (a) if = diagnosis is positive; terminate. (b) if > r diagnosis is n e g a t i v e ; terminate. (c) i f _< r go to step (with k = , n = r ) for further reduction. the set o f tables in table depicts the situation in a simplified form to make it easier to focus on the steps o f the reduction method. in realistic cases, actions may involve reports back to the user on the rules that are fired, auxiliary calculations (for instance, o f a statistical nature), or other options. in such cases, a table action portion would include additional information along with the number o f remaining tests that are depicted in this set o f tables. it should be noted that the use o f tables to describe the algorithm does not necessarily imply that implementation by tables is mandated. i f tables are used in the implementa- tion, they need not always be stored; that is, there are cases in which they can be generated. an example to illustrate the method, we use a specific example o f a / diagnostic problem. to solve the diagnostic problem: . pick any observations. . n a m e them temporarily t , t , t , and t . . check the first table in table with the results o f these observations: (p = positive or n = negative) . the possible cases are as follows: (a) i f the results o f all are p, then the diagnosis is definitely established. (b) i f only are p, then we need to check more o f the remaining . reduction algorithm in expert diagnosis s t a b l e . d e c i s i o n t a b l e s u s e d for /n, /n, /n, and /n p r o b l e m s t i t t t p p p p p p p p n n n n n n n n f p p p n n n n p p p p n n n n p p n n p p n n p p n n p p n n p n p n p n p n p n p n p n p n ~i f o r a n y /n p r o b l e m t p p p p n n n n t p p n n p p n n t p n p n p n p n f o r a n y /n p r o b l e m t i p p n n t p n p n f o r a n y /n p r o b l e m t i p n f o r a n y /n p r o b l e m (c) i f o n l y a r e p, t h e n w e n e e d to c h e c k m o r e o f the r e m a i n i n g . (d) i f o n l y is p, t h e n w e n e e d to c h e c k m o r e o f the r e m a i n i n g . (e) i f all a r e n , t h e n h y p o t h e t i c a l l y w e n e e d to c h e c k o f the r e m a i n i n g , w h i c h is i m p o s s i b l e ; thus, w e r e j e c t the d i a g n o s i s . in c a s e (a) t h e r e a r e o b s e r v a t i o n s ; all o f t h e m h o l d , and the d i a g n o s i s is e s t a b l i s h e d ( f u r t h e r c h e c k s a r e o f ). in c a s e s (b), (c), o r (d), a d i a g n o s i s is not akram salah and kevin d. reilly established because there is insufficient input information. instead o f restarting the problem, we can define a new reduced problem such that we check only the remaining observations. n o w we need to check either / in case (b), / in case (c), or / in case (d). in case (e) we can reject the diagnosis because the total number o f observations is , o f them have already been checked and failed, and the remaining observations are in number. to establish a diagnosis, observations need to exist; therefore, it is impossible to establish a diagnosis from this situation ( / ). thus, the method either establishes a diagnosis from the information provided or uses the information to reduce the problem to a smaller problem o f the same nature. the new problem can be solved recursively by the same methodology. commentary we can cite several advantages o f the reduction methodology: ( ) there is guaranteed recursive reduction until a solution is reached; ( ) the number o f rules to be checked is less than using the simple rule approach (see table ); ( ) the tables given in table can be used for any diagnostic problem k/n, regardless o f the value for n; and ( ) the growth o f the number o f rules is limited, as all the decision tables are complete (welland [ ]), and thus there is no possibility o f adding rules to any o f them. as can be seen, the n u m b e r o f rules in the reduction method depends on the length o f the subset that establishes the diagnosis, k, rather than the length o f the domain o f observations, n. this is an important property o f the reduction methodology, as in simple rule approaches the number o f rules grows exponentially with the length o f the set o f observations, assuming that simple rule generation uses either combinations or permutations. according to weiss and kulikowski ([ ], pp. - ), a problem similar to what we have been intimating was detected while an expert system for diagnosis for rheumatic diseases was being implemented. as the expert system evolved, the number o f its (physician) users increased; consequently, the number o f observations known to the system increased. the expert system started with a / diagnostic problem and was extended to / and eventually to / . a t a b l e . n u m b e r o f rules used to build a knowledge base using using using problem id permutation combination reduction / / / reduction algorithm in expert diagnosis simple treatment o f the p r o b l e m (see table ) would make the number o f rules increase exponentially with any increase in the number o f observations. a final point to be noted about the reduction method is that it does not depend on any particular application. the algorithm was developed for an expert system for differential diagnosis o f rheumatic diseases, but it can be used in any other rule representation o f the same nature. e n v i r o n m e n t the system that we e m p l o y for representing the methodology o f this article is based on an extension o f a previously defined system called e x p r d (extended prolog rule data system), an integration o f a prolog, a relational database, and a decision table system (salah [ ]). this system is used to store or generate decision tables such as those appearing here. prolog programs expressing the reduction algorithm are added as a part o f the management program. sets o f observations are stored in the system as database relations. an interactive dialogue prompts the user to provide the proper information for the diagnostic problem and invokes the reduction algorithm. i f a decision is reached, the program advises the user whether the diagnosis is established or rejected. i f the information is not sufficient to establish the diagnosis, the program prompts the user to provide m o r e information. an example dialogue in a session is as follows: give me a test you performed: arthralgia what is the result o f arthralgia ( p = positive, n = negative): p **diagnosis is p o s i t i v e * * **chronic p o l y a r t h r i t i s > weeks is a significant factor** what is the result o f synovial fluid inflammatory (p = positive, n = negative): p these results are not sufficient to establish a diagnosis akram salah and kevin d. reilly what is the result o f subcutaneous nodules (p = positive, n = negative): p • • • • *diagnosis is negative** • *two positive symptoms are noted** do you wish a trace o f this dialog? no • • • a general philosophy in dealing with rule systems emerges from our methodology: a management system is employed in which rules are dealt with as one o f the components. such a management system can be viewed as a meta-rule program that helps a user (or an expert-system administrator) to build, manipulate, query, and analyze a rule system. c o n c l u s i o n although the reduction methodology for differential diagnosis expert systems is self-contained in the sense that it solves a well-defined problem, if we take a broader view o f the situation, we see this method as part o f the overall rule- management environment. the environment conceptualization emphasizes use of meta-level processing to manipulate rule-like representations. given a k / n diagnostic problem, an extended form o f rule, the management program is designed to generate a set o f simple rules or employ the reduction methodology to reduce the problem to a smaller problem of the same nature. employing management programs on the meta-level facilitates a global .view for expert systems, allowing operations such as generation, reduction, or analysis o f rules. r e f e r e n c e s . codasyl decision table task group, a modern appraisal o f decision tables, association for computing machinery, new york, . . dahl, v., logic programming as a representation of knowledge, ieee computer, ( ), - , . . fikes, r., and kehler, t., the role of frame-based representation in reasoning, comm. o f the assn. f o r computing machinery , - , . reduction algorithm in expert diagnosis . salah, a., and yang, c. c., rule-based systems: a set-theoretic approach, proceedings o f the rd annual computer science symposium on knowledge- based systems: theory and application, columbia, sc, . . hayes-roth, f., rule-based systems, comm. o f the assn. f o r computing machinery , - , . . reilly, k., salah, a., morgan, p., and rowe, p., multiple representations in a language-driven memory model, in papers on computational and cognitive science (e. battistella, ed.), indiana univ. linguistic club, bloomington, ind., - , . . yang, c. c., relational databases, prentice-hall, englewood cliffs, n.j., . . reilly, k. d., salah, a., and yang, c. c., a logic programming perspective on decision table theory and practice, university of alabama at birmingham tech. report, . . kowalski, r., logic f o r problem solving, elsevier-north holland, new york, . . clocksin, w., and mellish, c., programming in prolog, springer-verlag, new york, . . bruynooghe, m., prolog in c f o r unix version : a reference manual, katholieke univ., leuven, belgium, . . prade, h., a computational approach to approximate and plausible reasoning with applications to expert systems, ieee transactions on pattern analysis and machine intelligence pami- , - , . . rosenbloom, p., laird, j., mcdermott, j., newell, a., and orciuch, e., r -soar: an experiment in knowledge-intensive programming in a problem-solving architec- ture, ieee transactions on pattern analysis and machine intelligence pami- , - , . . weiss, s., and kulikowski, c., a practical guide to designing expert systems, rowman & allanheld, philadelphia, . . welland, r., decision tables and computer programming, heyden & son, london, . . salah, a., a n integration o f decision tables and a relational database system into a prolog environment, phd thesis, univ. of alabama at birmingham, birmingham, ala., . nasa technical reports server (ntrs) mining the data from a hyperheuristic approach using associative classification this is a repository copy of mining the data from a hyperheuristic approach using associative classification. white rose research online url for this paper: http://eprints.whiterose.ac.uk/ / version: submitted version article: thabtah, fadi a. and cowling, peter i. orcid.org/ - - - ( ) mining the data from a hyperheuristic approach using associative classification. expert systems with applications. pp. - . issn - https://doi.org/ . /j.eswa. . . eprints@whiterose.ac.uk https://eprints.whiterose.ac.uk/ reuse items deposited in white rose research online are protected by copyright, with all rights reserved unless indicated otherwise. they may be downloaded and/or printed for private study, or other acts as permitted by national copyright laws. the publisher or other rights holders may allow further reproduction and re-use of the full text version. this is indicated by the licence information on the white rose research online record for the item. takedown if you consider content in white rose research online to be in breach of uk law, please notify us by emailing eprints@whiterose.ac.uk including the url of the record and the reason for the withdrawal request. mining the data from a hyperheuristic approach using associative classification fadi thabtah a,*, peter cowling b a department of computing and engineering, university of huddersfield, huddersfield, uk b mosaic research centre, university of bradford, bradford, uk abstract associative classification is a promising classification approach that utilises association rule mining to construct accurate classification models. in this paper, we investigate the potential of associative classifiers as well as other traditional classifiers such as decision trees and rule inducers in solutions (data sets) produced by a general-purpose optimisation heuristic called the hyperheuristic for a personnel scheduling problem. the hyperheuristic requires us to decide which of several simpler search neighbourhoods to apply at each step while constructing a solutions. after experimenting different solution generated by a hyperheuristic called peckish using different classifi- cation approaches, the results indicated that associative classification approach is the most applicable approach to such kind of problems with reference to accuracy. particularly, associative classification algorithms such as cba, mcar and mmac were able to predict the selection of low-level heuristics from the data sets more accurately than c . , ripper and part algorithms, respectively. � elsevier ltd. all rights reserved. keywords: associative classification; classification; data mining; hyperheuristic; scheduling . introduction heuristic and metaheuristic approaches have been applied widely in personnel-scheduling problems (blum & roli, ; cowling, kendall, & han, ). a metaheuris- tic could be defined as a recursive process which directs a simpler local search method by using different concepts for exploring and exploiting the search space in order to achieve good enough solutions (blum & roli, ). there are several different metaheuristic strategies for solving scheduling and optimisation problems such as local search, tabu search, simulated annealing and variable neighbour- hood search. hamiez and hao ( ) have used a tabu search-based method to solve the sport league scheduling problem (slsp). their implementation of the enhanced tabu search algorithm was able to schedule a timetable for up to teams and its performance in term of the cpu time was excellent if compared with previous algorithms such as that of (mcaloon, tretkoff, & wetzel, ) that had been used for solving the same problem. aicklen and dowsland ( ) have used genetic algorithms to deal with a nurse rostering problem in major uk hospi- tals, and hansen and mladenovic ( ) have showed that variable neighbourhood search is an effective approach for solving optimisation problems in which it generates good or sometimes near-optimal solutions in a moderate time. cowling, kendall, and soubeiga ( ) and cowling et al. ( ) argued that metaheuristic and heuristic approaches tend to be knowledge rich and require exten- sive experience in the problem domain and the selected heuristic techniques, and therefore they are expensive in term of their implementation. a new general framework to deal with large and complex optimisation and schedul- ing problems, called a hyperheuristic, has been proposed - /$ - see front matter � elsevier ltd. all rights reserved. doi: . /j.eswa. . . * corresponding author. e-mail addresses: f.thabtah@hud.ac.uk (f. thabtah), p.i.cowling@- bradford.ac.uk (p. cowling). www.elsevier.com/locate/eswa available online at www.sciencedirect.com expert systems with applications ( ) – expert systems with applications by cowling et al. ( ). it tends to robustly find good solutions for large and complex scheduling problems and has been proven to be effective in many experiments (cowling & chakhlevitch, ; cowling et al., ). a hyperheuristic approach can be described as a super- visor, which controls the choice of which local search neighbourhood to choose while constructing a solution/ schedule. a local search neighbour, also known as a low- level heuristic, is a rule or a simple method that generally yields a small change in the schedule. often these low-level heuristics are based on normal methods of constructing a schedule such as adding an event, deleting an event or swapping two events. fig. represents the general hype- rheuristic framework that at each iteration, selects and applies the low-level heuristic that has the largest improve- ment on the objective function, i.e. llh in the figure shown below. the arrows going from and to the hyper- heuristyic in fig. represent the selected low-level heuris- tics improvement values on the objective function obtained after trying them by the hyperheuristic. the training scheduling problem that we consider in this paper is a complex optimisation problem for a large finan- cial service company (cowling et al., ). it involves a number of events, trainers, and locations to be scheduled over a period of time. the task is to create a timetable of courses distributed over several locations in a specific period of time using a known number of trainers. a more detailed description of the problem is presented in the next section. in this paper, our aim is to determine an applicable data mining technique to the problem of deciding which low-level heuristic to apply in a given situation, using infor- mation about heuristic performance derived earlier. in particular, we would like to answer questions like: which learning algorithm can derive knowledge that could direct the search in order to produce good solutions? to achieve our goal, we compare three associative classi- fication techniques cba (liu, hsu, & ma, ), mcar (thabtah, cowling, & peng, ), mmac (thabtah, cowling, & peng, ) and two popular traditional classification techniques (part (frank & witten, ) and ripper (cohen, )) on data sets generated using a hybrid hyperhueristic called peckish (cowling & chakhlevitch, ), for the trainer scheduling problem. we analyse data from several solutions of the peckish hyperheuristic that combines greedy (best first) and ran- dom approaches. we identify that the learning task involves classifying low-level heuristics in terms whether they improved the objective function in old solutions in order to produce useful rules. these rules then will be used to decide the class attribute ‘‘low-level heuristic’’ while con- structing new solutions. we use the classification algo- rithms mentioned above to learn the rules. the training scheduling problem and different hyperheu- ristic approaches utilised to solve it are discussed in section . section is devoted to the applicability of data mining classification algorithms to predict the behaviour of low-level heuristics used by the peckish hyperheuristic. data sets, their features and experimental results are presented in section and finally conclusions are given in section . . the training scheduling problem and hyperheuristics a much simpler version of the training scheduling prob- lem has been solved in cowling et al. ( ) using a hyper- genetic algorithm. a larger and more complex problem, which has been described in cowling and chakhlevitch ( ) is summarised in this section. it involves a number of events, trainers, and locations to be scheduled over a period of time. the task is to create a timetable of geo- graphically distributed courses over a period of time using different trainers, and the aim is to maximise the total pri- ority of courses and to minimise the amount of travel for each trainer. the problem is associated with a large num- ber of constraints such as: • each event is to be scheduled at one location from the available number of locations. • each event must start within a specified time period. • each event can occur at most once. • each event to be delivered by competent trainers from the available trainers. • each location has a limited number of rooms and rooms have different capacities and capabilities. the data used to build the solutions of the training scheduling problem is real data provided by a financial firm where training is given by trainers over a period of months in different locations. further, there are about events to be scheduled and different low-level heuris- tics that can be used to build each solution. however, solu- tions given to us by the authors of cowling and chakhlevitch ( ) have been constructed using only low-level heuristics, where each low-level heuristic repre- sents local search neighbourhoods. for example, selecting a location with the lowest possible travel penalty for a par- ticular trainer to deliver a course as early as possible corre- sponds to a low-level heuristic. in solving the trainer scheduling problem, three hype- rheuristic approaches, random, greedy and hybrid, have been used. all of these approaches aim to manage the choice of which low-level heuristics to choose during the process of building the solution. the random approach llh llh llh llh llh llh hyperheurisic fig. . hypreheuristic general framework. f. thabtah, p. cowling / expert systems with applications ( ) – selects a low-level heuristic at random from the available ones in the problem. at each choice point in the search space, commonly all low-level heuristics have an equal chance to be selected. on the other hand, the greedy approach selects the low-level heuristic that yields the big- gest improvement in the objective function. if none of the available ones improve the objective function then the algorithm will be trapped in a local optimum. the hybrid approach is named peckish and consists of a combination of greedy and random approaches. it builds a solution by selecting a single low-level heuristic during each iteration in the search in order to apply. the choice is based on the low-level heuristic that has the largest improvement on the objective function in the problem (if one exists). in the case that none of the available low-level heuristics improve upon the objective function value, then the choice is random. in this paper, we choose a low-level heuristic from a can- didate list of good low-level heuristics. by changing the length of this candidate list and considering how it is merged, we can trade off the degree of greediness and ran- domness in the peckish hyperheuristic. as a result, several different solutions produced by the peckish hyperheuristic are investigated. we analysed the strategy used by the peckish hyperheu- ristic to construct a solution and observed that all available low-level heuristics in the problem must be tested in order to record their effect on the objective function at each iter- ation and apply only a single one. data mining could pro- vide a much quicker prediction of effective low-level heuristics at each iteration. in the next section, we investi- gate some of the popular data mining techniques for learn- ing the sets of low-level heuristics that improve the objective function and have been applied by the peckish hyperhueristic. . data mining for the selection of low-level heuristics since we are aiming to use knowledge derived from old solutions of the problem, data mining seems an appropri- ate technique to extract that knowledge. the next task is to identify which data mining method is applicable to extract knowledge from solutions generated by the peck- ish hyperheuristic. as mentioned earlier, the peckish hype- rheuristic usually selects and applies the low-level heuristic that leads to the largest improvement on the objective function and this is the class we want to find. in other words, we can learn rules that predict the performance of low-level heuristics in some solution runs and use these rules to forecast which low-level heuristics the hyperheu- ristic should choose in other runs. since we are predicting a particular attribute (low-level heuristic), as a result, supervised learning approaches such as classification are appropriate. there are many classification approaches for extracting knowledge from data that have been studied in the litera- ture cendrowska ( ), quinlan ( ) and cohen ( ). three common approaches, divide-and-conquer (quinlan, ), rule induction (cohen, ; furnkranz & widmer, ) and associative classification (li, han, & pei, ; liu et al., ) have been selected for our base comparison. further, five classification techniques related to such approaches have been compared, which are part (frank & witten, ), ripper (cohen, ), cba (liu et al., ), mcar (thabtah et al., ) and mmac (thabtah et al., ). our choice of these methods is based on the different schemes they use in learning rules from data sets. in the next subsection, we briefly survey these algorithms. . . associative classification associative classification techniques employ association rule discovery methods to find the rules. this approach was introduced in by ali, manganaris, and srikant ( ) to produce rules for describing relationships between attri- bute values and the class attribute and not for prediction, which is the ultimate goal for classification. in , asso- ciative classification has been successfully employed to build classification models (classifiers) by liu et al. ( ) and later attracted many researchers, e.g. (yin & han, ), from data mining and machine learning communi- ties. in this subsection we survey associative classification techniques used in this paper to generate rules from the hyperheuristic data. . . . classification based on association (cba) the idea of using association rule mining in classifica- tion problems was first introduced in liu et al. ( ), in which an algorithm called cba is proposed, which oper- ates in three main steps. firstly, if the intended data set contains any real or integer attributes, it is disctretised using multi-interval discretisation method of fayyad and irani ( ). secondly, the apriori candidate generation step (agrawal & srikant, ) is adopted to find the potential rules. apriori candidate generation method necessitates multiple passes, where the potential rules found in the previous pass are used for the generation of potential rules in the current pass. this repetitive scans requires high cpu time and main memory. once all poten- tial rules are produced, the subset that leads to the least error rate against the training data set is selected to from the classifier. the selection of such subset is accomplished using the database coverage heuristic, which ensures that every rule in the classifier must cover correctly at least one training data object. . . . mcar: multi-class classification based on association rule a recently developed associative classification algo- rithm called mcar (thabtah et al., ) employs tid- list intersections to quickly find the rule. this algorithm consists of two main phases: rules generation and a clas- sifier builder. in the first phase, the training data set is f. thabtah, p. cowling / expert systems with applications ( ) – scanned once to discover the potential rules of size one, and then mcar intersects the potential rules tid-lists of size one to find potential rules of size two and so forth. this rules discovery method does no require passing over the training data multiple times. in the second phase, rules created are used to build a classifier by considering their effectiveness on the training data set. potential rules that cover certain number of training objects will be kept in the final classifier. finally, mcar adds upon previous rule ranking approaches in associative classification, which are based on (confidence, support, rule length) by looking at the class distribution frequencies in the training data and prefers rules that are associated with dominant classes. experimental results showed that mcar rule ranking method reduces rule random selection during the process of ranking the rules especially for dense clas- sification data. . . . multi-class, multi-label associative classification (mmac) the mmac algorithm consists of three steps: rules gen- eration, recursive learning and classification. it passes over the training data in the first step to discover and generate a complete set of rules. training instances that are associated with the produced rules are discarded. in the second step, mmac proceeds to discover more rules that pass user pre- defined thresholds denoted by minimum-support and mini- mum-confidence from the remaining unclassified instances, until no further potential rules can be found. finally, rule sets derived during each iteration are merged to form a glo- bal multi-label classifier that then is tested against test data. the distinguishing feature of mmac is its ability of gener- ating rules with multiple classes from data sets where each of their data objects is associated with just a single class. this provides decision makers with useful knowledge dis- carded by other current associative classification algorithms. . . traditional classification approaches . . . c . c . algorithm was created by quinlan ( ) as a deci- sion tree method for extracting rules from a data set. c . is an extension of the id algorithm (quinlan, ), which accounts for missing values, continuous attributes and pruning of decision trees. as for the id algorithm, c . uses information gain to select the root attribute. the algo- rithm selects a root attribute from the ones available in the training data set. c . makes the selection of the root based on the most informative attribute and the process of selecting an attribute is repeated recursively at the so- called child nodes of the root, excluding the attributes that have been chosen before, until the remaining training data objects can not be split any more (quinlan, ). at that point, a decision tree is outputted where each node corre- sponds to an attribute and each arc to a possible value of that attribute. each path from the root node to any give leaf in the tree corresponds to a rule. one of the major extensions of the id algorithm that c . proposed is that of pruning. two known pruning methods used by c . to simplify the decision trees constructed are sub-tree replace- ment and pessimistic error estimation (witten & frank, ). . . . repeated incremental pruning to produce error reduction algorithm (ripper) ripper is a rule induction algorithm that has been developed in by cohen ( ). it builds the rules set as follows: the training data set is divided into two sets, a pruning set and a growing set. ripper constructs the classifier using these two sets by repeatedly inserting rules starting from an empty rule set. the rule-growing algorithm starts with an empty rule, and heuristically adds one condi- tion at a time until the rule has no error rate on the growing set. in fact, ripper is a refined version of an earlier developed algorithm called incremental reduced error pruning (irep) (furnkranz & widmer, ) that adds a post pruning heuristic on the rules. this heuristic has been applied to the classifier produced by irep as an optimisation phase, aiming to simplify the rule set fea- tures. for each rule ri in the rule set, two alternative rules are built; the replacement of ri and the revision of ri. the replacement of ri is created by growing an empty rule r i and then pruning it in order to reduce the error rate of the rules set including r i on the pruning data set. the revi- sion of ri is constructed similarly except that the revision rule is built heuristically by adding one condition at a time to the original ri rather than to an empty rule. then the three rules are examined on the pruning data to select the rule with the least error rate. the integration of irep and the optimisation procedure forms the ripper algorithm. . . . part unlike the c . and ripper techniques that operate in two phases, the part algorithm generates rules one at a time by avoiding extensive pruning (frank & witten, ). the c . algorithm employs a divide-and-conquer approach, and the ripper algorithm uses rule induction approach to derive the rules. part combines both approaches to find and generate rules. it adopts rule induc- tion approach to generate a set of rules and uses divide- and-conquer to build partial decision trees. the way part builds and prunes a partial decision tree is similar to that of c . , but part avoids constructing a complete decision tree and builds partial decision trees. part differs from ripper in the way rules are created, where in part, each rule corresponds to the leaf with the largest coverage in the partial decision tree. on the other hand, ripper builds the rule in a greedy fashion, starting from an empty rule, it adds conditions, until the rule has no error rate and the process is repeated. missing values and pruning tech- niques are treated in the same way as c . . f. thabtah, p. cowling / expert systems with applications ( ) – . data and experimental results . . data sets and their features data from different solutions produced by peckish hyperheuristic for the training scheduling problem were provided by the authors of cowling and chakhlevitch ( ). each solution represents iterations of applied low-level heuristics and is given in a text file. twelve of the data files each represents only a single solution, whereas, each of the remaining four files represents ten combined solutions. ten different low-level heuristics (llh , llh , llh , llh , llh , llh , llh , llh , llh , llh ) have been used to produce each solution. each file consists of different attributes and instances. one iteration in a single solu- tion is shown in table where the bold row indicates that low-level heuristic number was applied by the peckish hyperheuristic because it has the largest improvement on objective function. in table , column llh represents the low-level tested, esm and rsm columns stand for event and resource selection methods, respectively which specify how an event is scheduled. se indicates the selected event number to be scheduled. ue reflects whether another event is a conflict with currently scheduled event, i.e. they share the same trainer, location, or timeslot. eid corresponds to the event identification number and r stands for rescheduled, which means, if swapping unscheduled event from the schedule with the selected event is possible, then, it is possible to reschedule back the removed event from the schedule. op, np, ope and npe correspond to old priority, new priority, old penalty and new penalty, respectively. the new priority and new penalty values represent the total pri- ority and penalty of the schedule after applying a low-level heuristic. the difference between the new priority and new penalty values gives the value of the objective function. imp stands for amount of improvement and represents the change in value on the objective function after a low- level heuristic has been applied. cpu is the time in search for the low-level heuristic to be applied and sc reflects whether or not the current schedule has been changed, i.e. whether the current low-level heuristic had any effect at all, (‘ ’ changed or ‘ ’ not changed). finally ap column indicates whether or not the current low-level heuristic has been applied (‘ ’ applied or ‘ ’ not applied) by the hyperheuristic. after analysing the data in each file, we identified that six attributes have some correlation to the class attribute (llh), which are (op, np, ope, npe, imp, ap). how- ever, we are interested to learn rules that represent useful sequence of applied low-level heuristics at different itera- tions and lead to improvement on the objective function. therefore, solutions generated by the peckish hyperheuris- tic have been filtered to retain certain iterations where improvements have occurred upon the objective function. furthermore, a pl/sql program has been designed and implemented to generate a new structure of each solution in order to enable the extraction of rules that represent the sequence of applied low-level heuristics. in other words, each training instance in the new structure should contain the low-level heuristics that improved the objective func- tion in the current iteration along with others applied in the previous iterations. specifically, in each solution run and for each iteration, we record low-level heuristics applied in the previous three iterations along with the ones that improved the objective function in the current iteration. table represents part of a solution run (data) gener- ated in the new structure after applying the pl/sql pro- gram on the initial solution features, where columns llh_ , llh_ and llh_ represent the low-level heu- ristics applied at the previous three iterations. column llh represents the current low-level heuristic that improved the objective function and column imp repre- sents the improvement on the objective function value. finally column apply represents whether or not the selected low-level heuristic has been applied by the hype- rheuristic. as shown in table , data generated by the hyperheuristic have multiple labels, since there could be more than one low-level heuristic that improve the objec- tive function at any give iteration. hence, each training instance in the scheduling data may associate with more than one class. table one iteration of the peckish hyperheuristic for the training scheduling problem llh esm rsm se ue eid r op np ope npe imp cpu sc ap � � , , . � � , , . � , , . � � � , , � � , , � � � , , � � � , , � � � , , � � � , , � � � , , f. thabtah, p. cowling / expert systems with applications ( ) – . . experimental results in this section, we describe the experiments to evaluate classification accuracy and rules features produced by dif- ferent rule learning algorithms from the optimisation data sets. we have performed a number of experiments using ten-fold cross validation on different data files derived by the peckish hyperheuristics for the trainer scheduling problem. the sizes of the data files after applying the pl/ sql program vary. there are data files, each contains – training instances, where each file represents only one single run of the peckish hyperheuristic. the remaining four data files contain – training instances and represent ten combined solutions. two popular traditional classification algorithms (part, c . ) and three associa- tive classification techniques (cba, mcar, mmac) have been compared in terms of accuracy. the experiments of part and c . have been conducted using weka software system (weka, ) and cba experiments have been performed using a vc++ version provided by the authors of cba ( ). finally mcar and mmac algorithms were implemented in java under windows xp on . ghz, ram machine. the relative prediction accuracy, which corresponds to the difference of the classification accuracy of cba, part and c . algorithms with respect to (mcar, mmac) are shown in figs. – , respectively. fig. for instance signifies the difference of the accuracy between cba classifiers rela- tive to those derived by (mcar, mmac) on the optimisation data sets. the relative prediction accuracy figures shown are computed using the formulae ðaccuracymcar�accuracycbaÞ accuracycba and ðaccuracymmac�accuracycbaÞ accuracycba for cba. the same sort of formulae has been used for part and c . algorithms, respectively. we used a minsupp of % and a minconf of % in the experiments of cba, mcar and mmac algorithms. the label-weight evaluation mea- sure (thabtah et al., ) has been used to calculate the accuracy for mmac algorithm in the figures. - . % - . % - . % - . % . % . % . % . % . % . % data sets a c c u ra c y ( % ) d if fe re n c e m ca r m m a c cb a fig. . difference of accuracy between cba and (mcar, mmac). - . % - . % . % . % . % . % . % . % data sets a c c u ra c y ( % ) d if fe re n c e m ca r m m a c p a rt fig. . difference of accuracy between part and (mcar, mmac). table sample of optimisation data iteration llh llh llh llh imp apply f. thabtah, p. cowling / expert systems with applications ( ) – label-weight evaluation method assigns each class in a multi-label rule a value based on its number of occurrences with rule’s consequent (left-hand-side) in the training data. to clarify, a training object may belong to several classes where each one associated with it by a number of occur- rences in the training data set. each class can be assigned a weight according to how many times that class has been associated with the training object. thus, unlike the error- rate method (witten & frank, ), which considers only one class for each rule in computing the correct predictions, label-weight gives a value for each possible class in a rule according to its frequency in the training data. this gives the top ranked class in a rule the highest weight and not all the weight as error-rate method does. the accuracy results shown in the graphs indicated that associative classification algorithms outperformed the other learning techniques over the majority of test instances. particularly, cba, mcar and mmac outperformed the other learning algorithms on , , and benchmark problems, respectively. the won-loss-tied record of mcar against cba, c . and part are - - , - - and - - , respectively. the mmac won-loss-tied records against cba, c . and part are, - - , - - and - - , res- pectively. these figures show that associative classifica- tion approach is able to produce more accurate classifiers than decision trees and rule induction approaches, res- pectively. it should be noted that cba, part and c . algorithms outperformed (mcar, mmac) on data set number in figs. – , respectively. after analysing the data in this par- ticular set, it turned out that classes in this set are not evenly distributed. for example, class llh and llh were fre- quently applied by the hyperheuristic in this data set, whereas, classes llh , llh , llh and llh were rarely been used by the hyperheuristic. this is compound by the limited number of training instances in this particular data set ( training data objects). analysis of the classifiers produced revealed consistency in the accuracy of both part and c . because the differ- ence on average in the accuracy between them in all exper- iments is less than . %. this supports research works conducted in frank and witten ( ), where they show that despite the simplicity of part, it generates rules as accurately as c . and ripper. also c . and part algorithms showed consistency in the number of rules produced. analysis of the rules features generated from the hyper- huristic data has been carried out. fig. shows the number of rules extracted from nine data sets, categorised by the number of classes. mmac is able to extract rules that are associated with up to four classes for this data. this is one of the principle reasons for improving the accuracy within applications. fig. also demonstrates that the majority of rules created from each solution are associated with one or two class labels. it turns out that this reflects accurately the nature of the hyperheuristic data, since dur- ing each iteration, normally only one or two low-level heu- ristics improve on the objective function in the scheduling problem. thus, each training instance usually corresponds to just one or two classes. the additional classes discovered by mmac algorithm from the real data represent useful knowledge discarded by cba, part and ripper algorithms. the fact that - . % - . % - . % . % . % . % . % data sets a c c u ra c y ( % ) d if fe re n c e m ca r m m a c c . fig. . difference of accuracy between c . and (mcar, mmac). data set n u m b e r o f r u le s -label -labels -labels -labels fig. . distribution of the rules with regards to their labels. f. thabtah, p. cowling / expert systems with applications ( ) – mmac is able to extract rules with multiple classes enables domain experts to benefit from this additional useful infor- mation. in addition, these multi-label rules can contribute in the prediction step, possibly improving upon the classi- fication accuracy. the numbers of rules produced by c . , part, cba and mcar algorithms are listed in table . since mmac produces rules with multiple classes, we did not record their numbers of rules generated for fair comparison. the values in table show that c . always generates more rules than part, cba and mcar. this contradicts some earlier results reported in liu et al. ( ) and thabtah et al. ( ) on classifier sizes obtained against uci data collec- tion (merz & murphy, ), which show that associative classification approaches like cba and mcar normally generate more rules than decision trees. for this reason, we performed extensive analysis on the classifiers derived by c . from the optimisation data sets. after analysing the decision trees constructed by c . algorithm at each iteration, we observed that many rules are generated, which do not cover any training data. we found out that the reason for these many useless rules appear to be the attribute that c . splits the training instances on, if that attribute has many distinct values and only few of these values appear in the training data, then a rule for each branch will be generated, and hence only some of these branches cover training instances. the rest will represent rules that cover not even a single training instance. in other words, when the training data set consists of attributes that have several distinct values and a split occurs, the expected number of rules to be derived by c . can be large. since data sets used to derive the results in table contain four attributes where each one of them has different values (low-level heuris- tics) and some of these low-level heuristics are never result in solution improvement for the hyperheuristic, this explains the large numbers of rules derived by c . . . conclusions in this paper, we have studied data sets produced from a complex personnel scheduling problem, called the training scheduling problem. these data sets represent solutions generated by a general hybrid approach, called the peckish hyperheuristic, which is a robust and general-purpose opti- misation heuristic that requires us to decide which of sev- eral simpler low-level heuristic techniques to apply at each step while building the schedule. our study focused on analysing the behaviour of low-level heuristics that were selected by the hyperheuristic and improved upon the qual- ity of the current solution in order to extract useful rules. these rules can be used later to quickly predict the appro- priate low-level heuristics to call next. for this purpose, we have compared five data mining classification algorithms, (part, c . , cba, mcar, mmac) on different solu- tions produced by peckish hyperheuristic. the experimental tests showed a better performance for associative classification techniques (mcar, mmac, cba) over decision trees (c . ), rule induction (ripper) and part algorithm with reference to the accuracy of pre- dicting the appropriate set of low-level heuristics. since the mmac algorithm was able to produce rules with multiple classes, including very useful information that the hype- rheuristic can use in forecasting the behaviour of low-level heuristic while constructing a new solution. it is the most applicable data mining approach, which can be used to pre- dict low-level heuristic performance within the peckish hyperheuristic. furthermore, c . generated more rules than the other rule learning algorithms since useless rules were extracted by the c . algorithm, which have not even a single repre- sentation in the training data. the reason of these useless rules turn out to be the training data attributes, in which when some of these attributes are associated with many distinct values and only a subset of these values have suf- ficient representation in the training data, there will be valid rules for this subset and the rest will represent useless rules. references agrawal, r., & srikant, r. ( ). fast algorithms for mining association rule. in proceedings of the th international conference on very large data bases (pp. – ). aicklen, u., & dowsland, k. ( ). exploiting problem structure in a genetic approach to a nurse rostering problem. journal of scheduling, , – . ali, k., manganaris, s., & srikant, r. ( ). partial classification using association rules. in d. heckerman, h. mannila, d. pregibon, & r. uthurusamy (eds.), proceedings of the third international conference on knowledge discovery and data mining (pp. – ). blum, c., & roli, a. ( ). metaheuristics in combinatorial optimisa- tion: overview and conceptual comparison. acm computing surveys (csur) (pp. – ). cba ( ). . cendrowska, j. ( ). prism: an algorithm for inducing modular rules. international journal of man-machine studies, ( ), – . table number of rules of c . , part, cba and mcar on the optimisation data sets data c . part cba mcar f. thabtah, p. cowling / expert systems with applications ( ) – cohen, w. ( ). fast effective rule induction. in proceedings of the th international conference on machine learning (pp. – ). ca: morgan kaufmann. cowling, p., & chakhlevitch, k. ( ). hyperheuristics for managing a large collection of low level heuristics to schedule personnel. in proceedings of ieee conference on evolutionary computation (pp. – ). canberra, australia. cowling, p., kendall, g., & han, l. ( ) an investigation of a hyperheuristic genetic algorithm applied to a trainer scheduling problem. in proceedings of congress on evolutionary computation (cec ) (pp. – ). hilton hawaiian village hotel, hono- lulu, hawaii, may – , isbn - - - . cowling, p., kendall, g., & soubeiga, e. ( ). a hyperheuristic approach to scheduling a sales summit. in proceedings of the third international conference of practice and theory of automated timetabling (patat ). lncs (vol. , pp. – ). springer. fayyad, u., & irani, k. ( ). multi—interval discretisation of contin- ues-valued attributes for classification learning. in proceedings ofi jcai (pp. – ). frank, e., & witten, i. ( ). generating accurate rule sets without global optimisation. in proceedings of the fifteenth international conference on machine learning (pp. – ). madison, wisconsin: morgan kaufmann. furnkranz, j., & widmer, g. ( ). incremental reduced error pruning. in machine learning: proceedings of the th annual conference. new brunswick, new jersey: morgan kaufmann. hamiez, j., & hao, j. ( ). solving the sports league scheduling problem with tabu search. lecture notes in artificial intelligence (vol. , pp. – ). springer. hansen, p., & mladenovic, n. ( ). variable neighbourhood search. computer and operations research, – . li, w., han, j., & pei, j. ( ). cmar: accurate and efficient classification based on multiple-class association rule. in proceedings of the icdm’ (pp. – ). ca: san jose. liu, b., hsu, w., & ma, y. ( ). integrating classification and association rule mining. in proceedings of the kdd (pp. – ). new york, ny. mcaloon, k., tretkoff, c., & wetzel, g., . sports league scheduling. in proceedings of the third ilog optimisation suite international users conference. paris, france, . merz, c., & murphy, p. ( ). uci repository of machine learning databases. irvine, ca: university of california, department of information and computer science. quinlan, j. ( ). discovering rules from large collections of examples: a case study. in d. michie (ed.), expert systems in the micro-electronic age (pp. – ). edinburgh: edinburgh university press. quinlan, j. ( ). generating production rules from decision trees. in proceedings of the th international joint conferences on artificial intelligence (pp. – ). ca: morgan kaufmann. quinlan, j. ( ). c . : programs for machine learning. san mateo, ca: morgan kaufman. thabtah, f., cowling, p., & peng, y. ( ). mmac: a new multi-class, multi-label associative classification approach. in proceedings of the fourth ieee international conference on data mining (icdm ’ ) (pp. – ) brighton, uk. (nominated for the best paper award). thabtah, f., cowling, p., & peng, y. ( ). mcar: multi-class classification based on association rule approach. in proceeding of the rd ieee international conference on computer systems and applications (pp. – ). cairo, egypt. weka ( ): data mining software in java: . witten, i., & frank, e. ( ). data mining: practical machine learning tools and techniques association rules. in proceedings of the rd kdd conference (pp. – ). yin, x., & han, j. ( ) cpar: classification based on predictive association rule. in proceedings of the sdm (pp. – ). san francisco, ca. f. thabtah, p. cowling / expert systems with applications ( ) – mining the data from a hyperheuristic approach using associative classification introduction the training scheduling problem and hyperheuristics data mining for the selection of low-level heuristics associative classification classification based on association (cba) mcar: multi-class classification based on association rule multi-class, multi-label associative classification (mmac) traditional classification approaches c . repeated incremental pruning to produce error reduction algorithm (ripper) part data and experimental results data sets and their features experimental results conclusions references pii: s - ( ) - artificial intelligence ( ) – representation of propositional expert systems as partial functions robert m. colomb department of computer science and electrical engineering, the university of queensland, queensland , australia received february ; received in revised form december abstract propositional expert systems classify cases, and can be built in several different forms, including production rules, decision tables and decision trees. these forms are inter-translatable, but the translations are much larger than the originals, often unmanageably large. in this paper a method of controlling the size problem is demonstrated, based on induced partial functional dependencies, which makes the translations practical in a principled way. the set of dependencies can also be used to filter cases to be classified, eliminating spurious cases, and cases for which the classification is likely to be of doubtful validity.  elsevier science b.v. all rights reserved. keywords: propositional systems; machine learning; decision tables; decision trees; knowledge filtering . introduction there is a large class of expert systems whose purpose is essentially to classify cases, for example, to diagnose disease from symptoms. several different methods are used to represent the knowledge in such systems, and to form the basis of computer implementations, including propositional production rules, decision tables and decision trees. it has been known for some time that these representations are equivalent, in that a system represented in one can be automatically translated into another, exactly preserving the input–output behaviour of the original representation. this issue is addressed in detail by colomb and chung [ ]. translatability means that a body of knowledge can be represented in different ways for different purposes. for example, an expert system may be email: colomb@csee.uq.edu.au. - / /$ – see front matter  elsevier science b.v. all rights reserved. pii: s - ( ) - r.m.colomb / artificial intelligence ( ) – built by experts as a set of rules, which might be a natural way for humans to understand the knowledge; but a decision table representation can make it much easier to perform an analysis of the vulnerability of the knowledge base to measurement errors [ ]; and a decision tree representation may give a fast bounded time implementation [ , ]. on the other hand, an expert system may be built originally as a decision tree, say by induction from a set of cases [ , ], but be translated into one of the other forms for analysis or understandability. since these various representations of knowledge are interchangeable, we will call them all decision objects. however, even though the translations can be automatically performed by simple algorithms, there is a severe practical difficulty, namely that the translated representation can be very much larger than the original. furthermore, most of the constituent parts of the translated object may not be used in practice, even if all the parts of the original are. the main result reported in this paper is to show why this problem of greatly increased size occurs and to show a simple method of representing decision objects which eliminates it. to do this, we first exhibit a unified representation of the various forms of decision objects and present in this representation the translatability results from the literature. we then show how the translated representation becomes inflated, and that this inflation is the result of representing inherently partial functions on the attribute space as total functions. to solve the problem, we show a simple way of representing and estimating the domain of the partial classification function represented by the decision object, and show that by use of the knowledge of the domain, we obtain practical translation among forms of decision objects with little inflation. the paper concludes with a discussion and implications of the results. . forms of decision object a classification system typically processes a case consisting of measurements of a number of variables by assigning a classification to the case. this section is a collection and systematization of results from the literature. the proofs are therefore given somewhat informally since the full detail is available elsewhere. . . propositions we interpret a variable as a measuring instrument, used by a computer system to monitor a real-world process of some kind. a measurement by a specific variable is the assignment of a specific value to that variable, notionally by the real-world process. the value set belonging to a variable is a discrete set of names, usually describing qualitative properties. a value set must have at least two members. the prototypical case is a boolean variable with values {true,false}, but other value sets are possible: for example, the variable sex has the value set {male,female}. if a variable refers to a continuous measurement, its value set frequently names the results of a series of relational tests on the measurement: for example, the variable tsh might have the values high,borderline_high,normal,borderline_low,low,missing. there is no limit in principle r.m.colomb / artificial intelligence ( ) – to the cardinality of a value set. this notion of variable with finite value set is widely applicable: in particular, it is the basis for most expert systems. in an application, there are generally a number of variables. more formally, there is a set x={xi}, consisting of at least one variable. this set of variables is interpreted as a view of the real-world process. each variable xi has a value set vi. a state of the real-world process as measured by the system is an assignment xi =vj for vj in vi for all variables. this measurement of state is referred to as a case. the set x will be called the variable set associated with the system under consideration. an elementary proposition consists of a variable and a value from the value set belonging to that variable, designated xi = vij for vij in vi. we will designate by pi the set of elementary propositions pij associated with variable xi. the elementary propositions associated with the variables {xi} will be referred to as case elementary propositions. a case is represented as a proposition by a choice from {pi}, that is exactly one member of each pi. the ith choice will be designated ci and a case c will be represented by the set {ci}. each ci is a case elementary proposition associated with variable xi. classification is a function whose domain is the set of possible choices from {pi}, and whose range is a set of elementary propositions whose value set is a (typically small) finite set d={dj}. the classification assigned to a case will be represented as a proposition of the form class =dj for dj in d, where class is a reserved variable name. elementary propositions associated with the variable class will be referred to as classification elementary propositions. example . consider a system with two variables x (sex) with value set {female,male} and x (pregnant) with value set {true,false}. the case elementary propositions asso- ciated with each variable are p = {sex = female,sex = male} and p = {pregnant = true,pregnant = false}. there are four possible choices from {pi}, two of which are {sex = male,pregnant = false} and {sex = male,pregnant = true}. d is the set {normal,plausible,extraordinary}. assume that the classification assigned to the two cases exhibited are, respectively, the classification elementary propositions class = normal and class = extraordinary, while the classification assigned to the two other choices are both class= plausible. definition . an elementary proposition is binary if its value set has cardinality . proposition . the negation of an elementary proposition is a conjunction of elementary propositions. proof. an elementary proposition is a formula of the form xi =vij . if it is not true that xi =vij , then it must be true that xi =v, where v∈vi\vij , so that ∼(xi =vij)≡ ∨ k =j xi =vik, where k ranges over the indexes of the members of the value set vi. corollary . the negation of a binary elementary proposition is a binary elementary proposition. r.m.colomb / artificial intelligence ( ) – . . rules a rule system form of decision object is a collection of production rules. production rules are constructed from elementary propositions, classifications, and a distinguished set of propositions a={a} designated intermediate elementary propositions. intermediate elementary propositions are distinguished by having variables distinct from the variables which can occur in cases, and not including the reserved variable class. the antecedent of a rule is a conjunction of case or intermediate elementary propositions, or the negations of either, while the consequent is either an intermediate or classification elementary proposition. there is a single consequent in each production rule, so that a rule system is a collection of propositional horn clauses. a production rule system will be called well-formed if its dependency graph [ ] is acyclic. the dependency graph for a system of production rules has one node for each rule. an arc is defined with source n and target n if the consequent of the rule at n appears in the antecedent of the rule at n . note that this requirement is stronger than stratification as defined in the cited work. that work labels an arc “positive” or “negative” according to whether the consequent of n appears as a positive or, respectively, negative literal in the antecedent of n . a stratified system has no cycles including a negative arc, but allows cycles all of whose arcs are positive. the semantics of a propositional production rule based decision object is based on datalog under negation as failure. a set of propositional horn clauses is trivially a datalog intensional database (idb), since there are no variables. the extensional database (edb) consists of a set of predicates each of which corresponds to one of the elementary case propositions. each of the edb predicates either contains one tuple, which consists of the token “true”, or is empty. the standard stratified naive bottom-up evaluation procedure populates the idb predicates by propagating the “true” token. the classifications assigned to a case are the classification elementary propositions which are not empty. all this is a straightforward application of the principles of datalog as described, for example, by ullman [ , ]. example . we will re-cast example as a datalog system. the idb consists of the three horn clauses: (class = normal) :- (sex = male),(pregnant= false) (class = extraordinary) :- (sex = male),(pregnant= true) (class = plausible) :- (sex = female), and assume the edb represents the case of a pregnant male: (sex = male)={true} (sex = female)={} (pregnant= true)={true} (pregnant= false)={}. the datalog evaluation produces the following population of idb predicates (class = normal)={} (class = extraordinary)={true} (class = plausible)={}. r.m.colomb / artificial intelligence ( ) – the datalog semantics of rule systems shows that the well-formedness criterion which excludes recursion entirely involves no loss of generality. theorem . for every stratified propositional idb p there is an idb with an acyclic dependency graph which has the same perfect model. proof. based on the datalog naive evaluation procedure. this procedure starts with the given edb and no population for any idb predicate, and proceeds from stratum to stratum (making sure that clauses containing negated predicates in their antecedents are not evaluated until all the clauses with those predicates as consequents are fully evaluated). each step evaluates all clause bodies and possibly generates tuples for some of the idb predicates. this evaluation is called applying the t operator. the initial population is named t( ), and the population after n steps is called t(n). it terminates when a fixed point is reached: t(n+ )=t(n). t(n) is the perfect model. lemma . . in a propositional idb, each predicate generates tuples only once. proof. each predicate in the idb can have at most one tuple. therefore once a clause puts a tuple in a predicate, it is fully evaluated. even though a propositional predicate may be defined in several clauses, at most one clause firing fully evaluates the predicate. lemma . . for a given edb, the evaluation procedure selects an acyclic subset of idb clauses, which are the only clauses to generate tuples. proof. a clause can generate tuples only if all of its positive antecedent literals have populations, and none of its negative antecedent literals can possibly have populations. at each step n of the evaluation, include the clauses which generate tuples in t(n) in the required subset. designate the clauses added at t(n) by c(n), and the union up to n of the c(n) by s(n). lemma . insures that a clause can be in at most one c(n). that the required subset at the fixed point is acyclic can be seen by induction. clearly s( ) is acyclic, since the only clauses which can generate tuples in t( ) must have antecedents entirely in the edb. if s(n) is acyclic, then c(n+ ) includes only clauses all of whose antecedents contain only positive literals defined in s(n) and negative literals which are empty and which have been fully evaluated, so s(n+ ) can have no cycles. each edb e generates an acyclic subset of the idb by lemma . . designate that subset by s(e). lemma . . if e is an edb and s′ is an idb which generates no tuples when evaluated with edb e, there is an idb with an acyclic dependency graph which has the same perfect model as s′ when evaluated with edb other than e and the same perfect model as s(e) when evaluated with edb e. r.m.colomb / artificial intelligence ( ) – proof. by construction. if s(e) generated no tuples when evaluated with edb other than e, and s(e) had no intermediate elementary propositions in common with s′, then s′∪s(e) would be the desired idb. in general, this is not the case. we need to isolate s(e) from other edbs and to make sure all intermediate elementary propositions in s(e) do not occur in any other clause in the idb. it is easy to specify an additional set of rules which will serve to identify edbs. a propositional edb is a pattern of presence or absence of a token in a fixed collection of predicates. for edb e, we can create a rule with consequent i(e) which is true if the edb is the pattern e and empty otherwise. the antecedent of identification predicate i(e) has a positive literal for every proposition having a token in edb e, and a negative literal for every proposition lacking a token. also, since all elementary propositions assigned values in s(e) are assigned values by clauses in s(e) together with edb e, we can replace the names of the variables associated with all intermediate propositions in s(e) with names tagged by e. we will call this process isolation of an idb. for s(e), construct an idb s′(e) by including i(e) as a conjunct in the antecedent of all clauses in s(e). the only idb which can generate tuples when evaluated with edb e is s′(e). further, s′(e) has the same perfect model as s(e) (excluding the introduced identification predicates). if s′(e) is also the isolation of s(e), then s′ ∪s′(e) is idb required for the lemma. the identification predicate is what really does the job. the only reason for worrying about isolation is that the dependency graph is defined in terms of the propositions which are the consequents of clauses, not clauses in isolation. even though the introduction of the identification predicates is enough to keep s′(e) from any influence in the evaluation of any other edb, the isolation is required to separate the dependency graph. proof of theorem . construct the finite set (cardinality n) of all possible choices from {pi} as the set of possible cases c. impose an ordering on c, yielding a sequence of cases with the kth member identified by ck, k n. for each ck construct s(ck). construct s′′( ) as the empty idb. for each k > construct s′′(k) by the application of lemma . to s′′(k− ) and ck. s′′(n) is the required idb, which has an acyclic dependency graph and has the same perfect model as p for every possible case. theorem is of course not a practicable procedure to convert recursively defined propositional production rule sets to nonrecursive production rule sets. however, it shows that the use of recursion adds no expressive power. therefore it is reasonable to suggest to the programmer of a recursively defined propositional production rule set that the set be re-implemented to avoid recursion. definition . two decision objects are equivalent if they have the same perfect model for all edbs. r.m.colomb / artificial intelligence ( ) – . . decision table a decision table consists of a two-dimensional array of cells. associated with each row in the array is a classification. the ith cell in a row is a nonempty disjunction of elementary propositions from the set of elementary propositions associated with the ith variable. a row in a decision table can be viewed as a rule whose antecedent is a conjunction of cells, and whose consequent is a classification. therefore, a decision table can be viewed as a conjunction of row rules. example . the following is a decision table equivalent to the system in example . row sex pregnant classification sex= male pregnant = true class = extraordinary sex= male pregnant = false class = normal sex= female pregnant = true ∨ class = plausible pregnant = false definition . a cell is a don’t care cell if it is the disjunction of all the elementary propositions in the set associated with its column. the cell in example row associated with the variable pregnant is a don’t care cell. theorem . an acyclic rule system is equivalent to a decision table. proof. (detailed in [ ].) essentially done by successively unfolding (partially evaluating) the production rules, replacing the intermediate propositions by formulas consisting entirely of case elementary propositions or their negations. the result is a conjunction of rules whose consequents are classification elementary propositions. the result is expressed in clausal form. these clauses are rows in a decision table, with the exception that a given row may contain no elementary proposition from the set associated with a given variable. if this is the case, the antecedent of the clause is augmented by a conjunct consisting of the don’t care cell associated with the variable, which evaluates to true. definition . a decision table is unambiguous if no assignment of truth values to elementary propositions can assign true to more than one classification elementary proposition. an ambiguous decision table can arise from several sources. decision tables arising from theorem can be ambiguous due to errors in the rule set. one motivation for using theorem is to check the set of rules for ambiguity, which is very difficult to see in the rule representation. it can also be ambiguous only for impossible cases. insulating a system from impossible cases is one of the primary motivations of this paper. sometimes r.m.colomb / artificial intelligence ( ) – decision tables are used to identify situations in which there is normally one classification, but sometimes can have more-multiple disease states for instance. in this situation, we can without loss of generality replace the variables in the classification elementary propositions with cartesian products of variables, and the values by cartesian products of values, so that the form of the elementary proposition is preserved and the table becomes unambiguous, preserving the remainder of the present theory. definition . a decision table is complete if every assignment of truth values to elementary propositions assigns true to at least one classification elementary proposition. observation . a rule system can be equivalent to an ambiguous and/or incomplete decision table. see [ ]. . . decision tree a decision tree is a tree consisting of nodes of two classes: deciding nodes and leaf nodes. each deciding node is associated with a variable and each leaf node is associated with a classification elementary proposition. each arc has as its source a deciding node, and is associated with a case elementary proposition from the set associated with its source (associated arc proposition). our formulation is somewhat unusual in that each leaf node is associated not only with a classification elementary proposition but with a proposition composed from case elementary propositions. the operational semantics is that a leaf node fires on a case if all the arc propositions in a path to the root are consistent with the case, and so is the leaf proposition. a decision tree may therefore fail to classify a case. example . the following is a decision tree equivalent to the table in example . node sex—female—leaf (pregnant= true)∨ (pregnant= false)→ (class = plausible) | male node pregnant—true—leaf true → (class = extraordinary) | false leaf true → (class = normal). definition . a decision tree is complete if each deciding node is the source of arcs associated with all of the elementary propositions associated with its variable, and all of its leaf propositions are equivalent to true. definition . a path proposition is associated with a path from the root to a leaf, and is the conjunction of all the associated arc propositions conjoined with the leaf proposition. proposition . a conjunction of elementary propositions is either inconsistent or a choice from the sets of elementary propositions associated with a subset of the variables. proof. two distinct elementary propositions associated with the same variable are inconsistent. therefore, any consistent conjunction contains at most one elementary proposition associated with each variable. r.m.colomb / artificial intelligence ( ) – proposition . a path proposition is equivalent to a decision table. proof. the leaf proposition can be expressed in disjunctive normal form. the path proposition is therefore the conjunction of each of the leaf proposition disjuncts with all the arc propositions. by proposition , each of the resulting disjuncts contains at most one elementary proposition associated with each of the variables associated with the deciding nodes in the path and the variables in the path propositions, or is inconsistent. if inconsistent, the disjunct can be removed. finally, each disjunct can be augmented with don’t care cell propositions associated with any variable lacking any associated elementary proposition in the disjunct. theorem . a decision tree is equivalent to an unambiguous decision table, which is complete if the tree is. proof. each distinct path through the decision tree is equivalent to a decision table each row of which is associated with the path’s leaf classification elementary proposition. each path proposition is inconsistent with every other path proposition, since every pair of paths must share at least one deciding node and the arcs in each path whose source is that deciding node have different and therefore, inconsistent associated arc propositions (which are are case elementary propositions with the same variable and a different value for each arc). the union of the decision tables from all paths is a decision table, which is unambiguous, since each leaf and therefore, path has a single classification. if the decision tree is complete, then each deciding node partitions the set of possible cases, so there is a partition of the set of possible cases where each partition is associated with a leaf. the leaf proposition is true, so that every case in that partition is consistent with the decision table equivalent to the path proposition associated with the leaf. each case in the population of possible cases is therefore, contained in one of the parts of the partition, and is therefore, consistent with at least one row in the decision table. theorem . an unambiguous decision table is equivalent to a decision tree, which is complete if the table is. proof. by induction on the number of attributes associated with the table. we adopt the convention that a decision table with no attributes has a single row with the propositional value true, and therefore a single classification. basis step: if any of the following is the case, then the induction terminates with the indicated action: ( ) the number of attributes is zero. create a leaf node whose classification is the table’s single classification elementary proposition, and whose leaf proposition is true. (depends on the table being unambiguous.) ( ) the number of rows is zero. create a leaf node whose classification elementary proposition is arbitrary and whose leaf proposition is false. ( ) the number of distinct classifications is one. create a leaf node whose classification elementary proposition is the table’s unique classification, and whose leaf proposi- tion is the disjunction of propositions created from each row by the conjunction of the cell propositions. r.m.colomb / artificial intelligence ( ) – induction step: there is at least one attribute, at least one row, and at least two distinct classifications in the table t . ( ) select an attribute a by some method (this point is discussed below). ( ) create a deciding node associated with this attribute. ( ) for each v in the value set of a, create: • an arc whose source is the deciding node created and whose arc proposition is a=v. • a decision table t(a,v) derived by removing the cell associated with a from each row of t in which the cell associated with a is consistent with a=v. t(a,v) has one fewer attribute in its associated attribute set than t , so the induction proceeds. the tree is not complete unless every deciding node is the source of an arc associated with each elementary proposition associated with its variable. the arcs are created by part ( ) of the induction step. if one of the elementary propositions is missing, then the table has no row consistent with that elementary proposition and no case with that elementary proposition as a conjunct will be classified. similarly, if the leaf proposition does not evaluate to true, then no case consistent with the path proposition conjoined with the negation of the leaf proposition will be classified by the table, so the table must be incomplete. the tree is therefore, complete if the table is. from a practical point of view, the key to the algorithm developed from this theorem is in the method of choosing an attribute in part ( ) of the induction step. shwayder [ , ] uses a heuristic based on equalizing the number of rows in the t(a,v) (maximum dispersion), giving one of the standard algorithms for translating from a decision table to a decision tree. quinlan [ ] uses a heuristic based on minimizing the maximum number of distinct classifications in the t(a,v) (maximum entropy gain). theorem with the maximum entropy gain selection procedure gives a variant of the id algorithm. theorems and show that decision trees and unambiguous decision tables are exactly equivalent. . . cases quinlan’s paper [ ] is based on the concept of a training set of cases, and is intended to construct a decision tree from a training set. definition . a training set is a set of cases. associated with each case is a classification. proposition . a training set is a decision table. proof. by inspection. definition . a training set is noise-free if it is unambiguous. observation . theorem applied to a noise-free training set using the maximum entropy gain selection procedure is the id algorithm. r.m.colomb / artificial intelligence ( ) – . . ripple-down rule trees another form of propositional knowledge representation is the ripple-down rule tree [ ]. a ripple-down rule tree is a binary tree consisting of a set of decision nodes and a set of leaf nodes. a decision node is associated with a conjunction of elementary propositions. a leaf node is associated with a classification elementary proposition. there are two arcs whose source is a decision node. the true arc is taken if the node proposition is consistent with a case, the false arc otherwise. example . the following is an example of a ripple-down rule tree equivalent to the decision tree in example : node if (sex =male)&(pregnant= true)—true—leaf (class= extraordinary) | false node if (sex = female)—true—leaf (class= plausible) | false leaf (class = normal). definition . an rdr path proposition associated with a path through a ripple-down rule tree is constructed from a path from the root to a leaf. if the path takes a true arc, then the proposition associated with the decision node is a conjunct. if the path takes a false arc, then the negation of one of the conjuncts in the associated proposition is a conjunct. proposition . without loss of generality, an rdr path proposition is a decision table row. proof. an rdr path is a conjunction of disjuncts. each disjunct is a disjunction of elementary propositions associated with the same variable, since disjunctions are either single elementary propositions or are negations of elementary propositions (proposition ). by boolean algebra, the rdr path can be represented as a disjunction of conjunctions of elementary propositions. by proposition , each disjunct is either a choice from a set of variables or inconsistent. an rdr path is therefore, either inconsistent or a conjunction of disjuncts of elementary propositions associated with distinct variables. if the rdr path lacks elementary propositions associated with a given variable, the corresponding don’t care proposition can be conjoined. the classification is the classification associated with the leaf node. theorem . a ripple-down rule tree is equivalent to an unambiguous decision table. proof. the decision table is the union of all rdr path propositions. note that there can be many rdr path propositions associated with a single leaf, if the path from the root to the leaf takes any false paths whose source decision node is a nontrivial conjunction of elementary propositions, so there may be many rows in the decision table with the classification associated with a given leaf. however, rows with classifications from distinct leaves are inconsistent, since there must be a decision node in common, with one path taking the true arc and one the false arc. the table is therefore unambiguous. r.m.colomb / artificial intelligence ( ) – section has presented a framework in which the equivalences among the various forms of decision object (rules, decision tables, decision trees, cases and ripple-down rule trees) can be easily seen. using these equivalences, one can develop a classification system using a form of decision object convenient for the developers, then translate it into other forms for purposes of analysis and execution. . what is wrong with this picture? the idyllic theory described in the previous section encounters problems when put into practice. section describes the problems encountered and analyses their cause. some terminology is needed. we will say that a decision object has a number of parts. what a part is depends on the form of decision object. for a set of rules, a part is a rule; for a decision table, a row; for a decision tree or ripple-down rule tree, a leaf. the size of an object is the number of parts it has. a decision object is designed to solve a practical classification problem. we would expect that when we build a decision object, no matter what its size, that all of its parts will be used in practice. a part that is not used represents a cost giving no benefit. we will say that a decision object is compact to the extent that all of its parts are used in practice (measured, for example, by monitoring the decision object over a long period of use). generally speaking a decision object in its constructed form will be compact. the problem we encounter is that when an object is transformed, the new object can be very much larger than the original, and much less compact. we designate this phenomenon as inflation. for example, the transformations from theorems , and were applied to a well-known expert system garvan es [ ], which constructed clinical interpretations of thyroid hormone assays in a hospital pathology laboratory. it was in clinical use from about – , and was applied to many thousands of cases per year. the compactness of the various forms was tested using a sample of cases, representing more than one year’s use towards the end of the system’s life when its accuracy was more than %. garvan es was constructed as a set of production rules, of which at the end of its life there were , all of which were exercised by the set of cases. the system was, as one might expect, very compact. applying theorem , the production rules were transformed into an equivalent decision table which had rows. only of these rows were fired by any of the year’s cases. the transformed object is much larger than the original, and much less compact. theorem was then applied to the -row subset of the transformed decision table consisting of the rows which were fired by any of the year’s cases, using the maximum dispersion selection procedure, resulting in a decision tree with , leaf nodes, of which only were fired by any of the cases. the inflation is much worse than in the previous situation. finally, compton and jansen [ ] produced a ripple-down rule version of the knowledge base (called garvan rdr) which had leaves in its ripple-down rule tree, all of which were necessarily exercised in practice since the tree was constructed in a process based on the sample of cases. application of the algorithm derived from theorem would result r.m.colomb / artificial intelligence ( ) – in a decision table with , , rows. one would expect, of course, that some of these rows would be inconsistent or would be subsumed by other rows. nevertheless, only of the rows fire on any of the cases, giving an extreme inflation. inspection of the proofs of the theorems shows how the transformed object gets bigger than the original. theorem translates a set of production rules to a decision table by progressively substituting the antecedents of intermediate propositions for their occurrences in other rules. if the antecedent is a disjunction, or appears negated in a rule, or both, a rule absorbing the intermediate proposition gets broken into several rules whose antecedents are conjunctions of elementary propositions. if the absorbing rule has itself an intermediate proposition as a consequent, the rules multiply. the algorithm for transforming rules to decision tables derived from theorem therefore produces a result whose size is exponential in the length of a chain of intermediate propositions. (garvan es had an average of three intermediate propositions between the measurements and the classifications.) translating from a decision table to a decision tree, part ( ) of the induction step of theorem duplicates rows of the decision table where the cell associated with the selected variable contains a disjunction of elementary propositions. the algorithm derived from the theorem therefore, produces a result whose size is exponential in the number of variables. (garvan es had variables, and the decision tree had an average path length from the root to a leaf of about .) finally, the number of rows in the decision table equivalent of a ripple-down rule tree is more than the product of the number of conjuncts in the propositions associated with the decision nodes in the longest chain of false arcs. so we can see how inflation happens, but it remains to see why it happens. after all, the original decision object is created compact. we can see how it gets large, but where do all the unused parts come from? to see this, we step back to the black box view of a classification expert system. the system can be seen as monitoring some real-world process which generates cases. the system observes the cases through its variables. each distinct case is represented by a choice of values for the variables. the variables can therefore be seen to determine a space defined by the distinct possible choices. we call this space the attribute space for the process. it depends solely on the variables and their value sets. the attribute space for a process can be quite large. a convenient measure of its size is the number of bits necessary to represent a case, disregarding the statistics of the process. if all the variables are boolean, the size of the space is the number of variables. otherwise, the size of the space is the total of the log to the base of the cardinalities of the value sets. for example, garvan es has variables, whose value sets range in cardinality from to . its size is bits. if garvan es has an attribute space of bits, there are ≈ possible distinct cases. at cases per year and if all cases are different, it would take years before all the cases were encountered. in practice, as one might expect, there are fewer than distinct cases in the year’s history, some of which occur more than times. where the attribute space is large, one would expect that the process has many variables which are contingent on others taking specific values (pregnancy is a relevant variable only if the sex of the patient is female, for example). also, values of some variables occur in patterns r.m.colomb / artificial intelligence ( ) – depending on the values of others (the hormone profile of a pregnant woman or a child is different from a male adult, for example, and there are different characteristic disease states). in general, where the attribute space is large one would expect that almost every combination of values would produce a case the domain experts would regard as absurd. even if there were possible valid cases for the garvan es process, one would have to generate cases at random before the expert would find one that makes sense. this observation suggests that in systems with large attribute spaces the real-world process is confined to a very small region. the knowledge of the experts is confined to this small region, which we might call the region of experience. the expert system classifies cases, so can be seen as a function taking the region of experience to a space of classifications. this accounts for the compactness of the decision objects constructed. these decision objects are, however, constructed as total functions on the attribute space, not as partial functions. the strategy works essentially because by definition the process being monitored never generates an impossible case. however, this strategy leaves the expert system vulnerable to errors or maliciousness. the famous experiment in which lenat answered mycin’s questions as for a dead person, and mycin happily diagnosed a specific course of treatment, is almost certainly a case where the input lay outside mycin’s region of experience. decision objects get their compactness essentially by liberal use of don’t care propositions, or by conjunctions of elementary propositions which fail in limited ways. these elements are found and expanded by the transformation algorithms. inflation is a side effect of the practice of constructing classification systems as total functions. . a cure for inflation if inflation is a side effect of the construction of classification systems as total functions on large attribute spaces, then it would seem reasonable to look for a cure by defining the systems as partial functions. if one could get a characterisation k of the region of experience associated with a process generating cases to be classified, then we could use k to prune the transformed decision objects as they are being created. in theorem (rule → table), we would keep only those disjuncts of the definition of an intermediate proposition which are consistent with k. in theorem (table → tree), part ( ) of the induction step would only make copies of rows which are consistent with k. in theorem (rdr tree → table), we would keep only those rdr paths which are consistent with k. a good estimate of k would not only control inflation, but would act as a filter detecting cases which are either spurious or perhaps legitimate but outside the experience of the domain experts. a natural way to represent k is by a set of constraints stating that the values of certain variables are determined by the values of others. these sort of constraints are partial functional dependencies (pfds). the variables are in general independent, but if the set of variables in the domain of the dependency take on a particular combination of values, then the value of the variable in the range of the dependency is fixed. for example, in general the variables sex and pregnant are independent. however, if sex takes the value male, then r.m.colomb / artificial intelligence ( ) – pregnant takes the value false. similarly, if pregnant takes the value true, then sex takes the value female (at least in the experience of garvan es ). partial functional dependencies could be constructed by the domain experts as part of the process of building a decision object. certainly, the sex/pregnant situation in the previous paragraph is pretty straightforward. however, many expert systems are constructed by induction using various methods to produce various forms of decision object. if a body of cases is available, then it is plausible to induce a set of partial functional dependencies. in fact, one can see a priori that it must be possible to induce a set of partial functional dependencies. if the set of cases available is much smaller than the attribute space, then by assumption there are very many combinations of variable values which do not occur. the problem is therefore, not whether we can induce a set of pfds, but whether we can induce a set of practical size which gives a sufficiently tight upper bound on the region of experience, whether the induction can be done at an acceptable cost, and whether the set induced is statistically reliable. there are a number of algorithms available for estimating sets of pfds. one is from the method of rough sets [ ] which looks for rules resulting from value reducts, taking each attribute in turn as a decision attribute (set of classifications). another comes from the data mining literature [ ]. in data mining terminology, a pfd is an association with % confidence (no counterexamples). we want pfds with a small number of elementary propositions in their antecedents. there are two reasons for this. the first is that a single pfd with m elementary propositions in the antecedent and consequent taken together constrains the attribute space more the smaller m is. if all the variables are boolean and there are n variables, then a single constraint reduces the attribute space by a factor of n−m+ . the second reason for wanting pfds with few propositions in their antecedent is that the algorithms for computing pfds are exponential in the number of propositions in the antecedent. small pfds are therefore, both stronger and practical to compute. finally, we want pfds which are statistically reliable. in the absence of a good theory of the statistics of pfds, it seems reasonable to look for pfds which have a large number of positive examples (in data mining terminology, associations with high support). this prevents rare cases from contributing possibly spurious pfds. the existence of a good set of pfds is a question which can only be settled experimentally in particular situations. to gain experience, we estimated pfds for garvan es from the set of cases (of which are distinct), using the rough sets library. we applied the algorithm to the set of distinct cases, so that duplicated cases did not contribute to support. this process produced pfds with a single elementary proposition in their antecedent at a support level of and such pfds with a support level of . further, there were pfds with two elementary propositions in their antecedent at a support level of and such pfds at a support level of . since we are here concerned with the theoretical basis and computational feasibility of working with sets of pfds, but want statistically reliable constraint sets, we chose the upper support level ( examples with no counterexamples). we will call these two sets of constraints garvan k institute of computer science, warsaw university of technology, warsaw, poland. r.m.colomb / artificial intelligence ( ) – and garvan k , respectively, and in general a set of constraints with one antecedent k , and with two k . further experience might suggest a different support level, but at least in this case the number of rules computed does not vary greatly with differences in support level so it would make little difference in the following. this shows that it is possible, in at least one application, to induce a reasonable set of small pfds. the remaining question is how strong they are—the size of the region of experience defined by them, and what difference they make to the inflation problem. this requires some more theory. definition . a proposition covers a case if the case is consistent with the proposition. definition . a variable is missing from a proposition if the proposition contains no elementary proposition associated with it. one way to compute the strength is to use a method of cragun and stuedel [ ]. in our terminology, they show that if a proposition is in disjoint disjunctive normal form (each disjunct is inconsistent with every other) the number of cases covered by the proposition is the sum of the number of cases covered by each disjunct. the number of cases covered by each disjunct is one if the disjunct is the conjunction of propositions associated with all variables, and the product of the cardinality of the value sets of the missing attributes otherwise. unfortunately, this method is of limited utility, since k is constructed in conjunctive normal form, and the process of conversion from conjunctive to disjunctive normal forms is exponential in the number of conjuncts. garvan k requires the computation of ≈ conjuncts, while garvan k requires ≈ conjuncts, clearly beyond the bound of computational feasibility, even for garvan k only. one might think that there might be considerable redundancy in k, in that some of the constraints might be deduced from others. this is particularly easy to test in the case of k , since the redundancy can be viewed as a case of transitive closure. if a→b and b→c are both in k , then the algorithms used will ensure that the redundant a→c is also. applying this method to the -conjunct garvan k reduces it to conjuncts, taking the cost of conversion to disjunctive normal form from to , still well beyond computational feasibility. although there may be problems where the cragun and steudel approach is feasible, it is clear that there are some in which it is not. a monte carlo method based on constraint propagation is more generally applicable. the basis of the monte carlo method is the generation of random cases. a naive approach would generate a population of random cases and count the cases which satisfy k. this is not generally practical, since we are expecting the region of experience to be a tiny fraction of the entire attribute space—in the case of garvan es , the region of experience might be as small as − of the attribute space. if we augment the generation of random cases with propagation of constraints, we obtain a feasible method of generating an estimate of the size of the region of experience. r.m.colomb / artificial intelligence ( ) – algorithm . estimate the strength of a set of constraints k. the algorithm is based on generating cases by successively choosing random values for variables and propagating these values, until all variables are assigned values. the average number of random bits needed to generate values is an estimate of the size of the space covered by k. a variable which is not assigned a definite value will be referred to as an open variable. note that the method may restrict the possible values a variable may be assigned, but the variable remains open until a definite value is assigned (this last applies only to variables the cardinality of whose value set is greater than ). ( ) when sufficient cases have been generated terminate, returning the average number of bits needed to generate a case. ( ) generate a single case: • if all variables have values assigned, terminate, returning the number of bits needed to select values. otherwise: • select an open variable at random. generate sufficient random bits to give this variable a definite value, and assign that value. (note that the cardinality of the value set may be reduced due to earlier constraint propagation.) • propagate the new value to the other open variables using k. algorithm is a special case of the more general problem of propagation of finite domain constraints. it would be possible to build a constraint size estimating procedure using a finite domain constraint logic programming language [ ]. using an implementation of algorithm specifically designed for k constraint sets, garvan k was found to be of size . bits [ ]. since the attribute space for garvan es is of size bits, the region of experience for garvan k is − ≈ − of the total attribute space. this gain of bits is more than half of the maximum possible gain for garvan es . there are nearly distinct cases, so the region of experience is at least bits large, so the maximum gain is − = bits. this result suggests that a plausible strategy for inducing k is to first induce k , proceeding further only if k is not strong enough. a very simple algorithm suffices to compute k . algorithm . compute k from a set of cases c. construct an above the diagonal triangular array a of integers each of whose rows and columns is indexed by a member of one of the value sets of one of the variables underlying c. initialise this matrix to . for each member c of c, increment all the cells in a which correspond to pairs of values of variables in c. each cell of a now contains the number of instances the associated variable values co- occur in c. each variable value x =v selects a set of cells of a corresponding to all other variable values. if all but one of the cells associated with a given value is zero, with x′ =v′ indexing the nonzero cell, then x=v→x′ =v′ is a partial functional dependency. algorithm requires an array which is square in the aggregate cardinality of the value sets (which is about for garvan es ). it would be practicible to extend the algorithm r.m.colomb / artificial intelligence ( ) – to three dimensions to compute k , since the size of the corresponding matrix would be cubic in the aggregate cardinality of the value sets. we now know that it is possible, in at least one case, to economically induce a small, strong set of pfds. it remains to see what effect k has on the inflation problem. an experiment was run using the algorithm resulting from theorem (table → tree), using a slightly different attribute selection procedure, applied to the -rule decision table resulting from applying the algorithm derived from theorem to the garvan es rule set, then selecting only those rows which were consistent with all of the cases [ ]. in this experiment, the algorithm generated a tree with , leaves, of which were consistent with garvan k , and were consistent with at least one case. we can conclude from this section that the method of representing k as a set of partial functional dependencies is plausibly a practical approach to the inflation problem. the practicality issue is addressed in section below. . further properties of k the set k of constraints was introduced as a heuristic designed to control the inflation problem when translating propositional classification systems from one form to another. we observed in passing that another use for k was as a filter to detect cases that were either spurious or, if valid, at least outside the experience on which the classification expert system is based. if we take the use of k as a filter seriously, and stipulate that the expert system is a partial function defined only on the subset of the attribute space covered by k, we can incorporate k into the translation theorems obtaining a corresponding set of translation procedures which is likely to be much less susceptible to inflation. proposition . given a conjunction a of elementary propositions and a case c associated with the same variable set as a; either a is inconsistent, a and c are inconsistent, or a & c=c. proof. by proposition , a conjunction of elementary propositions is either inconsistent or a choice from the sets of elementary propositions associated with a subset of the variables. a case is a choice from the complete set of variables. therefore in the conjunction of an arbitrary consistent conjunction a of elementary propositions, for each elementary proposition in a there is an elementary proposition in c associated with the same variable. if they are distinct, a and c are inconsistent with each other. otherwise a & c=c by the boolean algebra identity a & a=a applied to the elementary propositions in the conjunction. proposition . for any consistent set k of partial functional dependencies and any case c covered by k, k & c=c. proof. k can in principle be represented in disjunctive normal form in elementary propositions. for each disjunct d, proposition gives either c&d = c or c is r.m.colomb / artificial intelligence ( ) – inconsistent with d. since c is consistent with k, at least one of the disjuncts must be consistent with c. the result follows from the boolean algebra identity a∨a=a. we now modify theorems , and . theorem (rules → table k). an acyclic rule system defined on the region of the attribute space covered by a set k of partial functional dependencies is equivalent to a decision table defined on the same region of the attribute space. proof. theorem proceeds by unfolding the set of rules, removing intermediate propositions, starting from intermediate propositions which are the consequents of rules having only elementary propositions in their antecedents. if a particular rule is of the form p→q, then the application of the rule to a case c follows the derivation process: ( ) c, p→q. ( ) c, c & p→q. ( ) if c is consistent with p: c,c→q (by proposition ). ( ) c, c→q, q. if q is an intermediate proposition, the process of unfolding in theorem creates a single proposition which is the only proposition whose consequent is the intermediate proposition, and similarly for its negation. by proposition , step ( ) in the derivation can be replaced by ( ′) c,c & k & p→q. if c is consistent with p, then k is consistent with p, for any c, by the definition of k. if the antecedent p is expressed in disjunctive normal form, then any disjunct inconsistent with k can be removed, as step ( ) of the derivation will never succeed. with this modification, the process of unfolding ultimately results in a decision table all of whose rows are consistent with k. an algorithm derived from theorem will prune the rows of the decision table as they are constructed, so will arrive smoothly at the final decision table consistent with k. theorems and were concerned with a complete propositional system, which was able to classify any possible case in the attribute space. completeness as defined in definitions and is no longer relevant, since we are looking at decision objects which are deliberately incomplete. this leads to the following definition of completeness with respect to a set of constraints. definition . a decision object p is complete with respect to a set of partial functional dependencies k if p classifies every case covered by k. proposition . if n(p) is the disjunction of cases not classified by p , then p is complete with respect to a set of partial functional dependencies k if n(p) is inconsistent with k. r.m.colomb / artificial intelligence ( ) – proof. by inspection. proposition may not lead to computationally feasible algorithms in all circumstances. however, if n(p) can be represented in disjunctive normal form with m conjuncts, then the test for competeness with respect to k is no worse than testing whether m cases are in the region of experience defined by k. one would expect that the leaf propositions in a decision tree would tend to be relatively small (note that don’t care cells can be eliminated by the boolean algebra identity a & true = a). a leaf proposition in disjunctive normal form with say conjuncts per disjunct, having say disjuncts, requires computation of conjuncts to convert its negation to disjunctive normal form, which is tolerable. most of these conjuncts would probably either be inconsistent or subsumed by others, so that the resulting expression would likely not have an excessive number of disjuncts. one would expect that there would be problems where the concept could be applied. definition leads to modifications of theorems and . theorem (tree → table k). a decision tree defined on the region of the attribute space covered by a set k of partial functional dependencies is equivalent to an unambiguous decision table defined on the same region of the attribute space, which is complete with respect to k if the tree is. proof. follows theorem , except that only paths through the decision tree consistent with k are included in the decision table, since if a path is inconsistent with k, no valid case will be consistent with it. if the tree is complete with respect to k, then for every case c covered by k, there is one path through the tree consistent with c. this path p(c) is consistent with k by definition of k, so is included in the table. thus every case c covered by k is consistent with at least one row of the table. theorem (table → tree k). an unambiguous decision table system defined on the region of the attribute space covered by a set k of partial functional dependencies is equivalent to a decision tree defined on the same region of the attribute space, which is complete with respect to k if the table is. proof. parallels that of theorem . replace part ( ) in the induction step with ( ′) for each v in the value set of a such that the path proposition in the tree down to a p(a,v) is consistent with k, create: • an arc whose source is the deciding node created and whose arc proposition is a=v. • a decision table t(a,v) with one row derived from each row r of t in which the cell associated with a is consistent with a = v, and such that p(a,v)& r consistent with k, by removing the cell associated with a. this proves that the tree is equivalent to the table on k, since no branch pruned is consistent with k, nor is any row excluded from the t(a,v). the tree is therefore complete with respect to k if the table is. finally, we modify theorem (rdr-tree → table), giving r.m.colomb / artificial intelligence ( ) – theorem (rdr-tree → table k). a ripple-down rule tree is equivalent with respect to k to an unambiguous decision table. proof. the decision table is the union of all rdr path propositions which are consistent with k. the remaining observations from the proof of theorem apply. we have now achieved a practical set of translation procedures for decision objects taking into account that they are partial functions on their attribute space. the issue of practicality is canvassed below. . discussion the main result of this paper has been the solution of the problem of inflation in the translation of decision objects from one form to another. the key idea has been the characterisation of the region of experience using a set of constraints expressed as partial functional dependencies. these problems are conceptually quite simple. their difficulty arises from their origin in exponential data structures and exponential algorithms. it is easy to construct worst-case examples where the computations are very long, so they depend for their utility on the typical case turning out to be tractable, in much the same way as the simplex algorithm in linear programming. the procedures have been tested on a large problem and shown to be practicable, but one would not be able to routinely recommend them until they had been used by a wide variety of people on a wide variety of problems and intractable cases did not arise in practice—again like the simplex algorithm—which is well beyond the scope of a single project. we can, however, consider the characteristics of situations where the procedures would fail. there are two things which could go wrong: • the set of pfds comprising k could have few rules with a small number of antecedents, instead very many rules with a large number of antecedents; or • the use of k could fail to prune the decision objects in the translation process. the former would be a problem both because the cost of computing k increases rapidly with the number of antecedents in the rules, and partly because of the cost of using a very large set of rules with a large number of antecedents. a system with an impracticable k would be extremely complex: it would be large, else the exponential algorithms would not reach their practical limits (a six-attribute decision table can be exhaustively analysed); and it would lack simple relationships so would be difficult for humans to either understand or to gather enough data for a statistically reliable decision object induction. the latter mode of failure requires that the inflation of the decision objects be highly anti- correlated with the set of constraints. this would be surprising, because the decision object is built from the relationships of the case elementary propositions to the classification elementary propositions and the set of constraints is built from the relationships of the case elementary propositions among themselves. one would think that a system where the two were strongly correlated would be noticeably peculiar. this highly informal argument indicates that the results of this paper are likely to be of wide applicability. r.m.colomb / artificial intelligence ( ) – solving the inflation problem with a set of constraints suggests a number of problems which will be the subject of further research. first, the study of the set of constraints itself: estimation of its size, simplifying it, etc., which can likely make use of the techniques of constraint logic programming (e.g., [ ]). second, results in the theory of propositional expert systems can be revisited and improved, for example, the computational stability of expert systems as reported in [ ]. in that work, the chief problem was to analyse a decision object in the form of a decision table to identify situations, where small changes in attribute space make large changes in classification space, using essentially a hamming distance measure on attribute space. the results could be improved by considering only errors which remain within the region of experience. of course, use of the set of constraints as a filter can improve the brittleness of a wide range of expert systems, improving results such as that of webb and wells [ ], which used method based on classification. in particular, there are problems such as described by davidsson [ ] in which a good characterisation of the region of experience is crucial. in that work, problems such as design of a coin-sorting machine are considered. in europe, one often finds coins from other countries in a set of coins to be sorted. it becomes essential to recognise that a coin is foreign and to reject it while sorting the coins of a particular country. use of partial functional dependencies should be able to improve davidsson’ results. further, in the collection of data describing the region of experience one could also obtain data describing a range of foreign coins, and so also have a set of cases which is explicitly outside the desired region of experience, which could improve the estimates of k. edwards et al. [ ] has investigated the problem of prudence in an expert system. the type of system concerned is an automated medical pathology laboratory system, which produces clinical interpretations of samples using the analysis machine results and a small amount of descriptive information about the patient. (the system, called piers, is a descendant of garvan es .) in the application, it is a legal requirement that all clinical interpretations be signed by an appropriate specialist. this requirement represents a significant residual human involvement. the proposal of edwards et al. is for the expert system to maintain records of all the cases it processes and to append to an interpretation how different the present case is from other cases previously processed. cases very similar to previous cases can be assumed to have very reliable interpretations, while the more different a case is from previous cases, the closer checking the interpretation should have. edwards et al.’s approach is to use the range of values for particular variables as a measure of similarity. we speculate that an incremental computation of a set of partial functional dependencies might give a more general and more reliable measure. in particular, we would investigate including all partial functional dependencies, no matter how weak their support, so long as they have % confidence. a new case can either increase the support of a dependency or break it, so that earlier estimates would tend to overconstrain the region of experience, which would be gradually widened as the system was used. each case violating a constraint would be subject to special scrutiny. finally, note that the characterisation of the region of experience by a set of constraints differs fundamentally from clustering methods: the set of constraints identifies the r.m.colomb / artificial intelligence ( ) – boundary of the region, while clustering methods generally identify the centres of regions of interest. it might be profitable to investigate the interaction of the two kinds of method. references [ ] r. agrawal, r. srikant, mining generalized association rules vldb’ , morgan kaufmann, los altos, ca, , pp. – . [ ] k.r. apt, h.a. blair, a. walker, towards a theory of declarative knowledge, in: j. minker (ed.), foundations of deductive database and logic programming, morgan kaufmann, los altos, ca, , pp. – . [ ] r.m. colomb, computational stability of expert systems, expert systems with applications ( ) – . [ ] r.m. colomb, y.-p. chen, use of partial functional dependencies to make practical approximate translations among forms of propositional expert systems, in: a. sattar (ed.), proc. th australian joint conference on artificial intelligence, lecture notes in artificial intelligence, vol. , springer, berlin, , pp. – . [ ] r.m. colomb, c.y. chung, strategies for building propositional expert systems, internat. j. intelligent systems ( ) – . [ ] r.m. colomb, j. sienkiewicz, analysis of redundancy in expert systems case data, in: proc. th australian joint conference on artificial intelligence, world scientific, singapore, , pp. – . [ ] p. compton, r. jansen, a philosophical basis for knowledge acquisition, knowledge acquisition ( ) – . [ ] b.j. cragun, h.j. steudel, a decision-table-based processor for checking completeness and consistency in rule-based expert systems, internat. j. man-machine studies ( ) – . [ ] p. davidsson, learning characteristic decision trees, in: proc. th australian joint conference on artificial intelligence, world scientific, singapore, , p. . [ ] g. edwards, b.h. kang, p. preston, p. compton, prudent expert systems with credentials: managing the expertise of decision support systems, internat. j. biomedical computing ( ) – . [ ] k.a. horn, p. compton, l. lazarus, j.r. quinlan, an expert computer system for the interpretation of thyroid assays in a clinical laboratory, australian computer j. ( ) – . [ ] j. jaffar, m.a. maher, constraint logic programming: a survey, j. logic programming / ( ) – . [ ] z. pawlak, rough sets: theoretical aspects of reasoning about knowledge, kluwer academic, dordrecht, . [ ] m.j. maher, constrained dependencies, theoret. comput. sci. ( ) – . [ ] j.r. quinlan, semi-autonomous acquisition of pattern based knowledge, in: j.e. hayes, d. michie, y.-h. pao (eds.), machine intelligence , ellis horwood, london, , pp. – . [ ] j.r. quinlan, induction of decision trees, machine learning ( ) – . [ ] t. sato, equivalence-preserving first order unfold/fold transformation systems, theoret. comput. sci. ( ) – . [ ] h. seki, unfold/fold transformations of stratified programs, in: g. levi, m. martelli (eds.), logic programming: proc. th international conference (lisbon), mit press, cambridge, ma, , pp. – . [ ] k. shwayder, conversion of limited-entry decision tables to computer programs—a proposed modification to pollack’s algorithm, comm. acm ( ) – . [ ] k. shwayder, extending the information theory approach to converting limited-entry decision tables to computer programs, comm. acm ( ) – . [ ] j.d. ullman, principles of database and knowledge-base systems, vol. , computer science press, rockville, md, . [ ] j.d. ullman, principles of database and knowledge-base systems, vol. , computer science press, rockville, md, . [ ] g.i. webb, j. wells, recent progress in machine-expert collaboration for knowledge acquisition, in: proc. th australian joint conference on artificial intelligence, world scientific, singapore, , pp. – . document downloaded from: this paper must be cited as: the final publication is available at copyright http://dx.doi.org/ . /j.eswa. . . http://hdl.handle.net/ / elsevier braysy, o.; martínez molada, e.; nagata, y.; soler fernández, d. ( ). the mixed capacitated general routing problem with turn penalties. expert systems with applications. ( ): - . doi: . /j.eswa. . . . the mixed capacitated general routing problem with turn penalties olli bräysy , eulalia mart́ınez , yuichi nagata , david soler ∗ ( ) agora innoroad laboratory, agora center, p.o.box , fi- university of jyväskylä, finland ( ) instituto universitario de matemática pura y aplicada. universidad politécnica de valencia camino de vera s/n, , valencia, spain ( ) interdisciplinary graduate school of science and engineering, tokyo institute of technology, nagatsuta midori-ku yokohama, kanagawa - , japan abstract in this paper we deal with the mixed capacitated general routing problem with turn penalties. this problem generalizes many important arc and node routing problems, and it takes into account turn penalties and forbidden turns, which are crucial in many real-life applications, such as mail delivery, waste collection and street maintenance operations. through a polynomial transfor- mation of the considered problem into a generalized vehicle routing problem, we suggest a new approach for solving this new problem by transforming it into an asymmetric capacitated vehicle routing problem. in this way, we can solve the new problem both optimally and heuristically using existing algorithms. a powerful memetic algorithm and a set of new benchmark instances are also ∗corresponding author: address: instituto universitario de matemática pura y aplicada. univer- sidad politécnica de valencia. camino de vera s/n, , valencia, spain. tel:+ - ; fax:+ . e-mail addresses: olli.m.p.braysy@jyu.fi (o. bräysy), eumarti@mat.upv.es (e. mart́ınez), nagata@fe.dis.titech.ac.jp (y. nagata), dsoler@mat.upv.es (d. soler). suggested. the experimental results show that the average deviation of the sug- gested solution method is less than . per cent with respect to optimum. keywords: vehicle routing, capacitated general routing problem, turn penalties, transformation. introduction in many real-life vehicle routing problems it is important to consider the risk and time associated with turns, i.e., the cost of the turn. moreover, some turns, especially u- turns, can be forbidden. this last implies that a vehicle route made with a classical graph route generator may be illegal if it does not respect the traffic signals. for example, figure shows a traffic signal located in an arterial avenue from valencia (spain). it indicates that the next two left turns are forbidden and the way to dodge them. these forbidden turns cannot be taken into account if we model the city map with a classical graph. considering turn penalties and forbidden turns is particularly important in down- town areas and for large-size vehicles. but also turn penalties are important in order to save time in tours on foot. from figure it is easy to see that going on foot, right turn a → b can be considered with zero cost, while turn a → c is much more time-consuming, because it implies to cross two streets, with up to two traffic lights. usually the real-life vehicle routing software are based on separate modules for shortest path calculation and vehicle route optimization. the latter is based on given time and distance matrices, calculated typically in the beginning with the shortest path procedure. in this context, the distance and time calculation are based on fixed and known stopping points (usually addresses) for the vehicles and the possible turn penalties and forbidden turns are taken into account during the shortest path calcu- lation. however, in practice, there are several applications where the exact stopping points are not known a priori and are part of the optimization problem, such as mail collection and delivery, waste collection and street maintenance operations. in these cases, including turn penalties and forbidden turns in the vehicle routing model is very important. figure . traffic signal indicating two forbidden left turns. figure . a street crossing. so far research on extended vehicle routing models with turn penalties and forbidden turns has been scarce. for previous research, see benavent and soler ( ), clossey et al ( ), corberán et al ( ) and soler et al ( ), that generalize several well- known single vehicle routing problems to the existence of turn penalties. they provide theoretical results about complexity and resolution and/or computational results on these extensions. the last cited work studies the most general problem, the mixed general routing problem (mgrp) with turn penalties, that includes all the cases studied in the previous cited papers. mgrp consists of finding a minimal cost closed walk on the links of a mixed graph g which traverses a given subset of “required” links and a given subset of “required” vertices. with respect to multivehicle routing problems, early papers that considered turn penalties focused on real-life applications and heuristic solution methods. see for exam- ple bodin et al ( ) and roy and rousseau ( ). later, turn penalties have been considered in the context of the mixed capacitated arc routing problem (mcarp), see for example bautista and pereira ( ) and belenguer et al ( ). the mcarp is an arc routing problem, in which a fleet of vehicles (with a known capacity) is based on a specified vertex (the depot) and must service a subset of the links of a mixed graph, with minimum total cost and such that the load assigned to each vehicle does not exceed its capacity and each link is serviced by exactly one vehicle. bautista et al ( ) present two ant colony metaheuristics for a real urban waste collection problem. this real problem is modeled as a particular case of the mcarp with turn penalties, in which they only consider two kind of turns: forbidden or allowed with zero cost. finally, perrier et al ( ) heuristically solve a real vehicle routing problem in the context of snow plowing operations that also takes into account the existence of forbidden turns. the multivehicle extension (with capacity constraints) of the mgrp is called the mixed capacitated general routing problem (mcgrp). due to its complexity, there are only few works on mcgrp, see e.g. jansen ( ) who studied the undirected case and pandit and muralidharan ( ). however, particular cases of the mcgrp, such as the capacitated arc routing problem (carp) and the capacitated vehicle routing problem (cvrp) have attracted a huge amount of research. in this paper we present a generalization of the mcgrp that considers turn penal- ties and forbidden turns. the objective is to minimize the sum of the costs of the traversed arcs and edges together with the penalties associated with the turns made. we call the new problem the mixed capacitated general routing problem with turn penalties (mcgrptp). as far as we know, this is the first time that mcgrptp is presented in the liter- ature. moreover, there is no previous research on multivehicle node routing problems with turn penalties. in this paper we also present for the first time an approach for solving capacitated routing problems with turn penalties, through suggesting a new polynomial transformation from the mcgrptp to an asymmetric cvrp (acvrp). to be more precise, the transformation is done in two steps: we first transform the mcgrptp into a generalized vrp (gvrp), using a new approach suggested in this paper. the key idea of gvrp, compared to cvrp is that in gvrp each customer has several alternative service locations, and only one of them has to be selected for service. for more details on gvrp, see e.g. ghiani and improta ( ). in the second step we transform the gvrp into an acvrp, using a model presented in soler et al ( ). finally, we present a set of new benchmark problems and a very powerful memetic algorithm for the acvrp, based on a previous study by nagata and bräysy ( ). the experimental results show an average deviation equal to . % for instances with known optimal solution and that large-size problems can be solved with the suggested ma. the suggested transformation makes it possible to use also any other powerful algorithm developed for acvrp, see e.g. fischetti et al ( ), vigo ( ) or the more recent heuristic by de franceschi et al ( ). the remainder of this paper is organized as follows. section introduces some def- initions and notations in order to formally define and solve the mcgrptp. in section , through two transformations, we prove that the mcgrptp can be transformed in polynomial time into a generalized vehicle routing problem (gvrp). it is known (soler et al ( )) that the gvrp can be transformed into an acvrp, so in section we show computational results for several sets of acvrp benchmarks. finally, in section we present our conclusions. definitions and notations first, to our aim, we formally define two known problems cited before: the asymmetric capacitated vehicle routing problem (acvrp) and the generalized vehicle routing problem (gvrp). the second one is an extension of the acvrp, introduced by ghiani and improta ( ), that can model several real-world situations and that will be the “cornerstone” to solve our problem: the acvrp is defined as follows: let g = (v, a) be a complete digraph, v = {vi}ni= being its set of vertices, where v is the depot vertex. each vertex vi with i > has an associated demand di > and each arc (vi, vj) ∈ a has an associated cost ci,j ≥ . moreover, a fleet of k vehicles with the same capacity w is available at the depot. find a set of k shortest routes, each starting and ending at the depot, such that each vertex vi (∀ i ∈ { , ..., n}) must be visited by one and only one vehicle and the sum of the demands of the vertices visited by each vehicle does not exceed its capacity w . as in other papers cited in the introduction, in all the capacitated routing problems discussed here, we will consider k to be equal to the minimum number of vehicles needed to serve all demands. the gvrp is defined as follows: let g = (v, a) be a directed graph where the vertex set v is partitioned into m + nonempty subsets s , s , ..., sm such that s has only one vertex v (the depot), sh (h = , ..., m) represents l(h) possible locations of the same vertex which has associated a positive demand dh, and each arc (vi, vj) ∈ a has associated a cost ci,j ≥ . moreover, a fleet of k homogeneous vehicles having the same capacity w is available at the depot. find a set of k shortest routes, each starting and ending at the depot, such that each subset sh (h = , ..., m) is visited exactly once and the sum of the demands of every route does not exceed the capacity w of the vehicle. next, we need to show some concepts and notations that have been used in previous works on turn penalties: given a mixed graph g = (v, e, a), each pair of links a = (u, v), b = (v, w) ∈ e ∪a has an associated turn at v, based on going from a to b, denoted as [ab]. moreover, if a, b ∈ e, the same pair has another associated turn at v, based on going from b to a, denoted as [ba]. each edge e incident with v has an associated u-turn at v that, if necessary, will be denoted by [eve]. each link in g has associated a nonnegative cost and each allowed turn in g has associated a nonnegative penalty. given a = (u, v), b = (s, t) ∈ e ∪a, a v-s feasible chain from a to b is an alternating sequence of links and allowed turns {a , [a a ], a , . . . , [ar− ar], ar, [arb]}, where a = a. the cost of a feasible chain is defined as the sum of the costs of the arcs it traverses plus the sum of the penalties of the turns it makes. a v-s feasible chain from a to b is closed if a = b and s = v. given a = (u, v), b = (s, t) ∈ e ∪ a, a shortest (minimum cost) v-s feasible chain from a to b will be denoted by s.f.c.(va, sb). note that a feasible chain is defined such that it begins at a link and ends at a turn. this is very important in the context of forbidden or penalized turns. in classical routing problems, if we have to go from a vertex u to a vertex v and then to a vertex w, we only have to connect the shortest path from u to v with the shortest path from v to w. but even if these shortest paths have been constructed taking into account turn conditions, the connection of both paths at v can give rise to an unavoidable forbidden turn (u-turn for example). in our case, the connection between two feasible chains at a vertex v is possible only if the first one ends at a turn [(t, v)(v, s)] and the second one begins at the link (v, s), which avoids the existence of forbidden turns. therefore, we cannot use paths between vertices as in the classical way, and this increases the difficulty of modeling how to serve demands at vertices, specially if these vertices have undesirable turns. with these previous concepts we can formally define the problem that we study in this paper. the mixed capacitated general routing problem with turn penalties (mcgrptp) is defined as follows: let g = (v, e, a) be a mixed graph where each link (i, j) ∈ e ∪a has an associated cost cij ≥ and each turn [ab] has an associated penalty p[ab] ≥ ( p[ab] = +∞ if turn [ab] is forbidden). one of the vertices, say v , represents the depot where there are k vehicles of an identical capacity w > and in it all the turns are allowed with zero penalty. let r ⊆ e ∪ a be a set of required links such that each (i, j) ∈ r has an associated positive demand qij ≤ w , and let vr ⊆ v −{v } be a set of required vertices such that each v ∈ vr has an associated positive demand qv ≤ w . find k closed feasible chains in g, one for each vehicle, that minimize the total cost and such that each chain passes through the depot, each demand is served by only one vehicle and the total demand served by each vehicle does not exceed its capacity w . note that allowing all the turns with zero penalties at the depot is due to the fact that in real-world situations, the depot normally represents a warehouse from which the vehicles begin their journey and to which they return. it makes no sense considering forbidden/penalty turns in the warehouse as the truck leaves from depot independently of the route the truck made before. moreover, these warehouses are usually placed outside the cities with good road communications and of easy access. hereinafter and as in similar papers, each non-required edge will be replaced by two arcs of the same cost and opposite direction; then we assume that all edges in the graph are required (er = e, with r = er ∪ ar). finally, for simplicity, we will not write the middle turns of a feasible chain. for example, the feasible chain {a, [ab], b, [b, c]} will be written as {a, b, [b, c]}. in the particular case of the mcgrptp in which vr = ∅, we have the mcarp with turn penalties, and in the particular case of the mcgrptp in which k = we have the mgrp with turn penalties. therefore, the problem presented here generalizes both the single vehicle and the multivehicle routing problems with turn penalties studied in the literature. solving the mcgrptp to solve the mcgrptp, we will first transform it into a gvrp, which in turn can be transformed into an acvrp. . transformation of the mcgrptp into a gvrp let g = (v, e, a) be a mixed graph where a mcgrptp is defined, e ∪ ar being the required link set and vr the required vertex set. due to the fact that in the gvrp the demand is at the vertices, we will transform graph g in which we have defined the mcgrptp into a directed graph g∗ = (v ∗, a∗) such that the vertices in g∗ are related with the required links and required vertices in g. to this aim, we first construct an intermediate directed graph g′ = (v ′, a′) from g as follows: first, each required edge is replaced in g by two opposite required arcs, both with the edge cost and the edge demand. second, subset vr is partitioned into two subsets, vr and vr , such that ver- tices containing all allowed zero-penalty turns belong to vr , and vertices containing forbidden or positive-penalty turns belong to vr . for each v ∈ vr ∪ {v }, replace vertex v in g by two vertices ve and vl, so that ve has only entering arcs (the arcs entering at v) and vl has only leaving arcs (the arcs leaving from v). add a required arc av = (v e, vl) to g such that all turns at ve and vl are allowed with zero penalty and the demand of this arc in g′ is the one of the required vertex v in g (the demand corresponding to the arc from the depot vertex is obviously zero). note that traversing arc av in g ′ is then equivalent to passing through vertex v in the original graph g. for each v ∈ vr do: - replace vertex v in g by the same number of vertices vij as those of allowed turns [aibj] at v, so that each of these copies has only one entering arc (ai) and one leaving arc (bj), with its corresponding allowed turn. note that if ai is an entering arc at v, g′ will contain at least as many copies of the entering arc ai as there are allowed turns involving ai at v. the same applies to the leaving arc bj from v. - then, replace each vertex vij by two vertices, v e ij and v l ij, and add a “required” arc avij = (v e ij, v l ij) between them with cost zero, such that p[aiavij ] = p[aibj ] and p[avij bj ] = , i.e. the penalty that was in the turn at vij is moved to vertex v e ij, and all these arcs will have the same demand, the one of the required vertex v. note that traversing only one of these required arcs avij in g ′ involves passing through vertex v in g. note that in the last paragraph we have written required in quotes because for each v ∈ vr , only one of the generated arcs must be served. after this transformation we have a directed graph g′ = (v ′, a′) such that the subset a′r comes from the required arcs, required edges and required vertices in g. a′r will give rise to a partitioned set of vertices v ∗ in the graph g∗ in which we will define the gvrp. each arc between two of these vertices that do not form part of the same subset, will have associated the cost of the shortest feasible chain between the two corresponding links in g′. from g′ we then construct the graph g∗ = (v ∗, a∗) as follows: - for each arc av ∈ a′r with v ∈ vr ∪ {v }, associate a vertex set sv = {xav} in g∗, with its corresponding demand in xav . - for each v ∈ vr , associate a vertex set sv in g∗ with as many vertices xavij as arcs avij are in a ′ r, all of them with the same corresponding demand. - for each arc a ∈ a′r that comes from a required arc in g, associate a vertex set sa in g ∗ with as many copies of a vertex xa as copies of arc a appear in g ′, all of them with the same corresponding demand. - for each pair of opposite required arcs e , e ∈ a′r that come from a required edge e in g, associate a vertex set se in g ∗ with as many copies of vertices xe and xe as copies of arcs e and e respectively appear in g ′, all of them with the same corresponding demand. - for each pair of vertices xa, xb ∈ v ∗ with xa ∈ si, xb ∈ sj, i = j being a = (u, v) and b = (s, t), add arcs (xa, xb) and (xb, xa) to a ∗, with the cost of the s.f.c.(va, sb) and of the s.f.c.(tb, ua) respectively in g′ (if they exist). - there is no arc between vertices belonging to the same si. given an mcgrptp in g, we define a gvrp in g∗ where the vertex set v ∗ is partitioned into the following subsets: sv for all v ∈ vr ∪ {v } (hereinafter we will denote the subset corresponding to the depot by s = {vd}), sv for all v ∈ vr , sa for all a ∈ ar and se for all e ∈ er. let us see an example of the construction of the auxiliary graph g∗. consider the mixed graph given in figure corresponding to an mcgrptp in which vertex represents the depot, there are two vehicles both with capacity , and there are three required arcs a, b and c, with demands , and respectively, a required edge e with demand and a required vertex with demand . then the total demand in the graph is units. link costs and demands appear in figure with different size (demands appear in bold and small size). we will suppose in this graph that all u-turns are forbidden except at vertex (the depot) at which all turns are allowed with penalty zero, and in the rest of the vertices, right turns (according to the drawing of the graph) have penalty , left turns have penalty except for the turn from arc b to arc ( , ) that is considered forbidden, and going straight ahead, as it occurs in the turn from arc a to edge e, has penalty zero. figure . graph g with ar = {a, b, c}, er = {e} and vr = { }. starting from the information given by the initial graph g = (v, e, a), where ar = {( , ), ( , ), (( , )}, er = {( , )}, vr = { } and the depot is vertex , we construct the intermediate graph g′ (see figure ): first, we replace the required edge e = ( , ) with demand qe = by two opposite arcs e , e with the same cost and the same demand . second, we replace vertex by the sequence { e, a , l}, a with cost zero. finally, we transform vertex , which belongs to vr (here vr = ∅), into four arcs, as many as allowed turns in it, all of them with cost zero and demand . figure . intermediate graph g′. from the intermediate graph g′ = (v ′, a′) we can already construct the directed graph g∗ = (v ∗, a∗). the vertex set v ∗ is given by the following partition: - a subset s with only one vertex xd representing the depot and corresponding to the required arc a . - subsets sa and sb, each one with only one vertex xa and xb respectively, for the required arcs a = ( , ) and b = ( , ), both with demand , and subset sc with two vertices xc and xc , both with demands , for the required arc c = ( , ) which have two copies in g′. - subset se with two vertices xe and xe both with demand , for the required edge e = ( , ). - subset s with four vertices x f h , x f c , x dc and x gh all of them with demand , for the required vertex in vr . figure shows graph g∗ in which, for simplicity, each pair of arcs (xr, xt) and (xt, xr) with xr ∈ si, xt ∈ sj and i = j has been drawn as a line with two arrow heads, one at each end, and the arc costs (normally different for each direction) have been omitted. figure . directed graph g∗ associated with g. theorem an mcgrptp defined in g can be transformed in polynomial time into the corresponding gvrp defined in g∗. proof. by the construction of g′, given b = {ti}ki= a set of k feasible closed chains in g corresponding to a solution to the mcgrptp, we can associate with b a set b′ = {t ′i}ki= of k feasible closed chains in g′ such that for all i ∈ { , . . . , k} we have: - t ′i traverses arc a (corresponding to the depot node in g). - ti passes through a vertex v ∈ vr iff t ′i traverses arc av in g′. - ti passes through a vertex v ∈ vr iff t ′i traverses an arc avij . - ti traverses arc a ∈ ar iff t ′i traverses a copy of arc a in g′. - ti traverses edge e ∈ e iff t ′i traverses a copy of arc e or a copy of arc e in g′. - t ′i has the same cost as ti. - moreover, we will suppose that if ti satisfies demand at v ∈ v (v ∈ v ) (a ∈ ar) (e ∈ e), then t ′i satisfies the demand located at av (one and only one arc avij ) (one and only one copy of arc a in g′) (one and only one copy of e or e in g ′). summarizing, for all i ∈ { , . . . , k} t ′i is a feasible closed chain in g′ that traverses a , satisfies the same demands as ti and has the same cost as ti. once we have constructed b′ from b, for each i ∈ { , . . . , k}, from t ′i we construct a cycle cbi in g ∗ as follows: suppose that t ′i satisfies, in this chronological order, the demands qj , qj , . . . , qjmi . then cbi is a cycle in g ∗ that visits, in this order, the sets of vertices s , sj , sj , . . . , sjmi , s , and for all t ∈ { , . . . , mi}, ci visits only the vertex at st coming from the arc in g′ at which demand has been satisfied by t ′i . it is evident that the set of cycles lb = {cbi }ki= is a solution to the gvrp in g∗ its cost c∗(lb) being less than or equal to c(b), this last due to the fact that the cost of the route segment of one feasible closed chain t ′i in g ′ between two consecutive serviced arcs (including the depot arc a ) is greater than or equal to the cost of the shortest feasible chain in g′ between them (from the first serviced one to the second serviced one), while the cost of the arc in cbi in g ∗ from the vertex coming from a serviced arc in g′ to the vertex coming from the next serviced arc in g′ is equal to the one of the shortest feasible chain in g′ between them. on the other hand, let l = {ci}ki= be a set of k cycles corresponding to a solution to the gvrp in g∗. for each i ∈ { , . . . , k}, from ci we construct a feasible closed chain t ′li in g ′ as follows: let (xa, xb) be a generic arc of ci, a = (u, v) and b = (w, r) being the arcs in a ′ r from which vertices xa and xb come, respectively. arc (xa, xb) will give rise in t ′l i to the s.f.c.(va, wb), and such that t ′li assumes the demand at a (the same as the one in xa) and b (the same as the one in xb). note that in this way, t ′l i has the same cost as ci and satisfies the same demands as ci. from the set b′l = {t ′li }ki= of k feasible closed chains in g′, we construct now a set bl = {t li }ki= of k feasible closed chains in g as follows: - “contract” each sequence in t ′li of the form (u, v e)(ve, vl)(vl, w) with av = (v e, vl) if v ∈ vr ∪ { } by (u, v)(v, w) in t li . - “contract” each sequence in t ′li of the form (u, v e ij)(v e ij, v l ij)(v l ij, w) with avij = (veij, v l ij) if v ∈ vr by (u, v)(v, w) in t li . - if t ′li traverses a copy of arc e or a copy of arc e in g ′, with e ∈ e, replace this copy in t li by edge e. - any other link or turn in t ′li is replaced by itself in t l i . - demand at v ∈ v (v ∈ v ) (a ∈ ar) (e ∈ e) is assigned to route t li iff t ′li satisfies the demand located at av (one and only one arc avij ) (one and only one copy of arc a in g′) (one and only one copy of e or e in g ′). it is evident that bl = {t li }ki= is a solution to the mcgrptp in g with c(bl) = c∗(l), this last due to the fact that for all i, c(t li ) = c ′(t ′li ). let then lopt be an optimal gvrp solution in g∗ and let bl opt be the mcgrptp solution in g obtained from lopt as described above. for each mcgrptp solution b in g we have: c(b) ≥ c∗(lb) ≥ c∗(lopt) = c(blopt ) ( ) therefore, bl opt is an optimal mcgrptp in g. . solving the gvrp in g∗ through an acvrp once we have transformed our mcgrptp into a gvrp defined in g∗ = (v ∗, a∗), from g∗ we construct a digraph ĝ = (v̂ , â) as follows: - v̂ = v ∗. - for each si with i ∈ { , , . . . , m} and |si| > , order its vertices consecutively in an arbitrary way {vi , . . . , vil(i)}; then, for j = , . . . , l(i) − , define the cost of arc (vij, v i j+ ) ∈ â as zero; also define the cost of arc (vil(i), vi ) as zero. - for every vij ∈ si and every w /∈ si define the cost of arc (vij, w) in â equal to the cost in g∗ of the arc (vij+ , w) ((v i , w) if j = l(i)) plus a fixed positive large quantity m if |si| > , and equal to the cost in g∗ of arc (vij, w) plus m if |si| = . - any other arc in â has infinite cost. - assign positive demands having sum equal to di to the vertices in si ∀ i, except for the depot subset. as it was proved by soler et al ( ), the gvrp can be transformed into an acvrp. following their proof, we can see that to solve the gvrp in g∗ we can solve the acvrp in the digraph ĝ, and to obtain a gvrp solution from an acvrp one, we just have to identify each path in the acvrp solution (vij, v i j+ , . . . , v i l(i), v i , . . . , v i j− , w) w /∈ si (w can be the depot) if j = and |si| > , or (vi , . . . , vil(i), w) if j = and |si| > , or (vij, w) if |si| = (in this last case vij can be the depot), with the arc (vij, w) in g ∗. an optimal acvrp solution hopt in ĝ, will give rise to an optimal gvrp solution lopt in g ∗ with cost c∗(lopt) = ĉ(hopt) − m (m + k). going on with our example, from the graph g∗ where the gvrp is defined (figure ) we define the acvrp in the digraph ĝ (see figure ) where, for simplicity again, the pairs of arcs (xr, xt) and (xt, xr) with xr ∈ si, xt ∈ sj and i = j have been drawn as lines with two arrow heads, one at each end, and the arc costs (normally different for each direction) have been omitted. figure shows the cost zero “intraset” arcs and the demand assigned to each vertex. for example, vertex , belonging to vr and with demand in g, has associated the set s in g ∗ with four vertices that in ĝ have demands , , and , respectively. figure . directed graph ĝ associated with g∗. figure shows the optimal solution h to the acvrp in ĝ given in figure corresponding to our example; it consists of two cycles: - h = (xd, xa, xb, xd) with associated cost + + + m = + m (these arc costs will be deduced above when explaining the solution to the mcgrptp in g) and the demand served by this cycle is qa + qb = + = . - h = (xd, x dc, x gh, x f h, x f c, xc , xc , xe , xe , xd) with associated cost: + + + + + + + + + m = + m and the demand serviced by this cycle is qc + qe + qv = + + + + + = . note that the total cost ĉ(h) = + m = + + ( + )m since m = is the number of vertex subsets in g∗ (the depot not included) and k = is the number of cycles in the solution. this optimal solution h in ĝ gives rise to the optimal solution l in g∗ shown in figure , with total cost c∗(l) = ĉ(h) − m = , and that consists of two cycles: c = (xd, xa, xb, xd) with cost and c = (xd, x dc, xc , xe , xd) with cost . figure . optimal solution to the acvrp in ĝ. figure . optimal solution to the gvrp in g∗. let us find the optimal solution to the mcgrptp in g given by these cycles, according to the proof of theorem . from c = (xd, xa, xb, xd) we have: - (xd, xa) corresponds to the s.f.c.( la , a) in g′: ( e, l)( l, )[( l, ), ( , )] with cost + + + = (in the addition, normal-font numbers indicate arc costs while bold numbers indicate turn penalties). see figure to follow the chains in g′. - (xa, xb) corresponds to the s.f.c.( a, b) in g′: ( , )[( , ), ( , )] with cost + = . - (xb, xd) corresponds to the s.f.c.( b, e a ) in g′: ( , )( , ef h)( e f h l f h)( l f h, )( , e) [( , e), ( e, l)] with cost + + + + + + + + + = . therefore, joining the chains we have the following feasible closed chain in g′ t ′ = { ( e, l), ( l, ), ( , ), ( , ), ( , ef h), ( e f h, l f h), ( l f h, ), ( , e), [( , e), ( e, l)] } with cost , and assuming the demands of arcs ( , ) and ( , ) (a and b), with a total demand of . from t ′ , following again with the proof of theorem , we construct the feasible closed chain with the same cost as t ′ in the original graph g, t = {( , ), ( , ), ( , ), ( , )( , ), ( , ), [( , ), ( , )]} that also assumes the de- mands of arcs a and b ( + ). similarly, from c = (xd, x dc, xc , xe , xd) we have: - (xd, x dc) corresponds to the s.f.c.( la , ea dc dc ) in g ′: ( e, l)( l, )( , edc)[( , e dc), ( e dc, l dc)] with cost + + + + + = . - (x dc, xc ) corresponds to the s.f.c.( la dc dc , lc dc ) in g ′: ( edc, l dc)[( e dc, l dc), ( l dc, )] with cost + = . - (xc , xe ) corresponds to the s.f.c.( c , e ) in g′: ( ldc, )[( l dc, ), ( , )] with cost + = . - (xe , xd) corresponds to the s.f.c.( e , e a ) in g′: ( , )( , e)[( , e), ( e, l)] with cost + + + = . therefore, joining the chains we have the following feasible tour in g′, t ′ = { ( e, l), ( l, ), ( , edc), ( e dc, l dc), ( l dc, ), ( , ), ( , e), [( , e), ( e, l)] } with cost , and assuming the demands of arcs ( edc, l dc), ( l dc, ) and ( , ), with a total demand of . from t ′ , we construct the feasible closed chain with the same cost as t ′ in g, t = {( , )( , )( , )( , )( , )[( , ), ( , )]}, that assumes the demands of vertex belonging to vr , of arc ( , ) and of edge ( , ) ( + + ). figure shows the optimal solution to the mcgrptp in the original graph g: t with solid bold line and t with broken bold line. figure . optimal solution to the mcgrptp in g. computational experiments the aim of this section is to show that the transformation presented here can be considered as a good tool to solve mcgrptp instances, at least heuristically due to the complexity of the problem. that is, if there exists any competitive procedure to solve the acvrp, we can solve mcgrptp instances within a reasonable running time. to do this, we first present a powerful heuristic algorithm for the acvrp based on the memetic algorithm (ma) for the (symmetric) cvrp proposed by nagata and bräysy ( ). ma is a population-based heuristic search approach that combines evolutionary algorithm with local search algorithm. although there are other heuristic approaches for the acvrp, as those cited in the introduction, we selected the above mentioned ma because it is shown to be currently the most powerful heuristic method for the cvrp, and it can be applied to the acvrp by a straightforward extension. here the suggested ma has been tested on a set of acvrp instances by pessoa et al ( ). we have also applied the ma to a set of single vehicle instances by soler et al ( ) because the optimal solutions are known to these instances. finally, the ma has been applied to a set of acvrp instances with up to vertices that come from the transformation of mcgrptp instances. . the ma for the acvrp the main feature of the ma by nagata and bräysy ( ) is that the edge assem- bly crossover (eax) operator generates offspring solutions by combining edges of two solutions selected as parents from the population. the generated offspring solutions may violate the capacity constraint. in this case, a subsequent local search-based re- pair procedure is used to restore the feasibility of the temporarily infeasible solutions. moreover, a simple local search is applied to the obtained feasible solutions according to a standard ma procedure. note that the eax was adapted to the acvrp by defining it on the directed graph whereas the original eax for the symmetric cvrp was defined on the undirected graph. the ma by nagata and bräysy minimized the total travel distance without putting any constraint on the number of vehicles and the number of vehicle was also a decision variable. however, according to the literature and therefore to the definition of the acvrp given here, in the acvrp the travel distance must be minimized with a given number of vehicles. so we have made a new version of the ma that minimizes the total travel distance for a fixed number of vehicles. the suggested ma has been implemented in c++ and has been executed on a adm opteron . ghz computer. for each instance, the ma has been executed five times. . results for the acvrp instances we first analyze the efficiency of the two ma versions (fixed or variable number of vehicles) on a set of acvrp instances with known upper bounds that appear in the work by pessoa et al ( ), which are variants of the benchmark acvrp instances given in http://or.ingce.unibo.it/research/cvrp-and-dcvrp. as far as we know, the work by pessoa et al is the most recent paper with computational results on the acvrp. the results for the acvrp instances without constraint on the number of vehi- cles are presented in table . the columns in the table list instance names (instance), the capacity of the vehicles (c), the number of vehicles and the total travel distance of the best-known upper bound solutions (k and u b), the number of vehicles and the total travel distance of the best result in five runs (best-k and best-d.) with our ma, the average number of vehicles and the average total travel distance in these five runs (ave-k and ave-d.), and the average computation time in seconds for a run (time). table . computational results on known acvrp instances. instance c k best u b best − k best − d ave − k ave − d t ime a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . a - f . . . we can see that in four instances out of the instances the ma has improved the best known upper bound by using one more vehicle, and in the other instances the solution given by the ma coincides with the best known solution both in the total travel distance and in the number of vehicles. we also have run the ma with fixing a priori the number of vehicles equal to the one given in the best known solution (see third column in table ). in this case, in all of the instances the total travel distance obtained with the ma coincides with the best known upper bound, with running time similar to the one given in table . therefore, the table corresponding to these results has been omitted. in table we show the results to the set of single vehicle instances given by soler et al ( ) for which the optimal solution is known. in this table, results are averaged over selected groups of the instances where the instances are partitioned into groups according to features of the instances. the columns in the table lists the number of instances in each groups (ins), the average number of required arcs in the subset (ara), the number of (required) edges in all the instances in the subset (|e|), the average number of required vertices in the subset (arv ), the average number of vertices in the asymmetric tsp instances obtained from the original instances (avat sp ), the average time in seconds to obtain the optimal solution with the exact procedure (at o), the average time in seconds to obtain the heuristic solution with the ma (at m a), the number of instances optimally solved with the ma in the subset (op t ), and the average deviation of the ma solutions in the subset (adev ). by deviation we mean (u b − lb)/lb. table . computational results on known single vehicle instances. group ins ara |e| arv avat sp at o at m a op t adev . , . , , , , , , , , , , , , . , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , , > h. , we can see from table that the results obtained with the ma are very good. the ma was able to optimally solve instances out of the instances, with . % average deviation. . solving mcgrptp instances with the ma we applied the ma to solve also three sets of random mcgrptp instances with up to arcs, (required) edges, vertices, required arcs, and required vertices, which are transformed into acvrp instances with up to vertices. as far as we know, the largest acvrp instance described in the literature until now has vertices. next we describe the data generation procedure for each set. the first set was generated from the single vehicle instances given by soler et al ( ), and we obtain two mcgrptp instance sets (one with vehicle capacity and the other with vehicle capacity ) as follows: we choose the depot node as the first required vertex in the numerical order. in this node all turns are changed to be allowed with zero cost. each required arc, required edge and required vertex will have a randomly generated integer demand in range [ , ], [ , ], and [ , ] respectively. note that when we transform one of these mcgrptp instances into an acvrp instance, if an original required vertex v has demand dv and it gives rise to a vertex subset in the acvrp with t vertices, let r be the largest integer positive number such that dv = rt + c, with c ≥ , then t − vertices in this subset will have demand r, and the last one will have demand r + c. these generated mcgrptp instances act as a basis for acvrp instances with number of vertices in the interval [ , ]. the second set contains mcgrptp instances generated from of the biggest single vehicle instances by corberán et al ( ). each original instance is first trans- formed into an mgrp with turn penalties instance with the procedure explained in soler et al ( ) and then transformed into two mcgrptp instances (one with ca- pacity and the other with capacity ) with the procedure given above for the first set. these instances form the acvrp instances with number of vertices in the interval [ , ]. finally, the third set contains instances that have been obtained from new large single vehicle instances randomly generated with the same instance generator used by corberán et al ( ). these mcgrptp instances are transformed into acvrp instances with the number of vertices in the interval [ , ]. in these three sets we have used the ma version with variable number of vehicles in order to obtain best upper bounds with respect to the total travel distance. the appendix shows a table containing all data and results corresponding to each individual mcgrptp instances. in this table, each instance is named as ixxy, where xx indicates the subset to which the instance belongs and y indicates the number of the instance inside that subset. the first subsets correspond to the first set of ( ) instances, subsets to correspond to the second set of ( ) instances and subsets to correspond to the third set of ( ) instances. the columns in the table list the following data: the number of vertices in the acvrp instance obtained from the corresponding mcgrptp instance (|vacv rp |), the total demand in the acvrp instance (t.d.), and for i ∈ { , }, the total travel distance of the best result in five runs obtained for a vehicle capacity of ci = i· units (di), the number of vehicles corresponding to this best result (ki) and the average computing time in seconds for a run corresponding to the capacity ci (timei). note that di is the total distance in the original mcgrptp instance, not in the auxiliary acvrp instance. the ma was able to find feasible solutions for all instances including the large- size instances with up to vertices, within a reasonable computation time, as re- ported in the appendix. the computation times vary from a few seconds to more than one hour depending on the problem size. we consider the reported computing times reasonable, given the size and complexity of the considered acvrp instances. based on these results, one can conclude that at least medium-size real-world mc- grptps can be solved by a state-of-the-art heuristic method for the acvrp through their transformation into an acvrp as explained here. by the way, we have gener- ated a large number of instances to the, until now, limited set of acvrp benchmarck instances. of course these new instances will be available to any researcher interested on them. conclusions in this paper we have studied a generalization of the mcgrp including turn penalties and forbidden turns. through an intermediate transformation into a gvrp, we have provided a procedure to transform it into an acvrp. then, at least from a theoretical point of view, this generalization can be solved both optimally and heuristically with existing algorithms. we have also introduced a set of new benchmark problems and adapted a recent and powerful memetic algorithm to acvrp. the experimental results show an average deviation equal to . % for instances with known optimal solution and that large-size problems can be solved with the memetic algorithm. we are convinced that research on turn penalties will increase and be of important value in the future to reduce the gap between theoretical research and real-life appli- cations. we hope that the theoretical and experimental results presented here can be used in the future as ideas or tools to test the efficiency of specific procedures to solve capacitated routing problems with turn penalties. acknowledgements this work has been partially supported by the ministerio de educación y ciencia of spain (project tin - -c - ). appendix inst. |v |-|a|-|e| |ar| |vr| |vacv rp | t.d. d k time d k time i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . inst. |v |-|a|-|e| |ar| |vr| |vacv rp | t.d. d k time d k time i - - . . i - - . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . inst. |v |-|a|-|e| |ar| |vr| |vacv rp | t.d. d k time d k time i - - . . i - - . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . inst. |v |-|a|-|e| |ar| |vr| |vacv rp | t.d. d k time d k time i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . i - - . . references bautista, j., & pereira, j. ( ). ant algorithms for urban waste collection routing. lecture notes in computer science, , - . bautista, j., fernández, e., & pereira, j. ( ). solving an urban waste collec- tion problem using ants heuristics. computers & operations research, , - . belenguer, j.m., benavent, e., lacomme, p., & prins, c. ( ). lower and upper bounds for the mixed capacitated arc routing problem. computers & operations research, , - . benavent, e., & soler, d. ( ). the directed rural postman problem with turn penalties. transportation science , - . bodin, l., fagin, g., welebny, r., & greenberg, j. ( ). the design of a com- puterized sanitation vehicle routing and scheduling for the town of oyster bay, new york. computers & operations research, , - . clossey, j., laporte, g., & soriano, p. ( ). solving arc routing problems with turn penalties. journal of the operational research society, , - . corberán, a., mart́ı, r., mart́ınez, e., & soler, d. ( ). the rural postman problem on mixed graphs with turn penalties. computers & operations research, , - . de franceschi, r., fischetti, m., & toth, p. ( ). a new ilp-based refinement heuristic for vehicle routing problems. mathematical programming series b, , - . fischetti, m., toth, p., & vigo, d. ( ). a branch-and-bound algorithm for the capacitated vehicle routing problem on directed graphs. operations research, , - . ghiani, g., & improta, g. ( ). an efficient transformation of the generalized vehicle routing problem. european journal of operational research, , - . jansen, k. ( ). bounds for the general capacitated routing problem. net- works, , - . nagata, y., & bräysy, o. ( ). edge assembly based memetic algorithm for the capacitated vehicle routing problem. networks, , - . pandit, r., & muralidharan, b. ( ). a capacitated general routing problem on mixed networks. computers & operations research, , - . perrier, n., langevin, a., & amaya, c.v. ( ). vehicle routing for urban snow plowing operations. transportation science, , - . pessoa, a., poggi, m., & uchoa,e. ( ). robust branch-cut-and-price algo- rithms for vehicle routing problems. presented at route (international work- shop on vehicle routing and transportation), jekyll island, georgia. roy, s., & rousseau, j.m. ( ). the capacitated canadian postman problem. infor, , - . soler, d., mart́ınez, e., & micó, j.c. ( ). a transformation for the mixed general routing problem with turn penalties. journal of the operational research society, , - . soler, d., albiach, j., & mart́ınez, e. ( ). a way to optimally solve a time- dependent vehicle routing problem with time windows. operations research letters, , - . vigo, d. ( ). a heuristic algorithm for the asymmetric capacitated vehicle routing problem. european journal of operational research, , - . international journal of engineering and advanced technology (ijeat) issn: – , volume- issue- , april retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication knowledge based expert system for predicting diabetic retinopathy using machine learning algorithms j.jayashree, sruthi r, ponnamanda venkata sairam, j.vijayashree abstract: diabetic retinopathy (dr) is a medical condition that can affect the patient's retina and cause leaks in the blood due to diabetes mellitus. the increase in cases of diabetes limits existing manual testing capability. today new algorithms are becoming very important for assisted diagnosis. effective diabetes diagnosis can benefit the victims and reduce the negative harmful effects, including blindness. if not treated in a timely manner, this disorder can cause different symptoms from mild vision problems to total blindness. early signs of dr are the hemorrhages, hard exudates, and micro-aneurysms (hem) that occur in the retina. timely diagnosis of hem is important for avoiding blindness this paper presents pso feature selection algorithms with three classifications for the detection of diabetic retinopathy using python. keywords: diabetic retinopathy, feature selection, classification, complications, treatment, prevention, statistics. i. introduction diabetic retinopathy is the undergoing changes that take place in blood sugar levels throughout the capillary of the retinal system. some vessels may swell up in some cases, and fluid leaks into the back of the eye. these could swell and drop in the capillary. or they could close, blocking the flow of blood. anomalous new capillary often grow up on the retina. all these improvements will rob your eyesight. dr was not leading symptoms initially, just low vision complications. ultimately, that it will lead blindness. whoever also type and type diabetic can develop the condition. the you have far more diabetes, and the less sugar on your blood regulated, the greater the probability that you will experience this eye complication. in other cases abnormal arteries will grow on the ground of the retina. over time, too much blood sugar will contribute to blocking the tiny capillary that feed the retina and sever blood supply. as a consequence, the eye looks for new capillary development a. types of diabetic retinopathy there are two types of diabetic!retinopathy: early diabetic!retinopathy commonly known as -non proliferative diabetic retinopathy (npdr) which occurs when there isn’t growth/proliferating of new capillary. revised manuscript received on april , . j. jayashree, school of computer science and engineering, vit,vellore vijayashree.j@vit.ac.in sruthi r, school of computer science and engineering, vit,vellore ponnamanda venkata sairam, school of computer science and engineering, vit,vellore j. vijayashree, school of computer science and engineering, vit,vellore that is the initial phase of diabetes eye disease. the walls of the capillary within the retina weaken when you have npdr. smaller bulges (microaneurysms) extend down from the walls of the smaller vessels, frequently withering fluid and blood into another retina. greater retinal shafts, too, may start dilating and radius is abnormal. npdr can switch from mild to severe, as more siege capillary. capillary within the eye can also close off with npdr. this is called ischemia macular. when this occurs, the macula cannot be penetrated by blood. occasionally, small particles, termed exudates, could perhaps form in the retina. nerve fibers in the retina can start swelling. central segment of the retina (macula) sometimes starts swelling (macular edema), an ailment which needs medical attention. advanced!diabetic retinopathy, known as proliferative diabetic!retinopathy, can progress to this more serious type. in this case, damaged capillary narrow off, brings new prosperity, irregular capillary of retina, and may dip into the transparent, fliud-like fluid that filling your (vitreous) middle of eye. pdr is the most advanced stage of eye disease for diabetics. it happens when new capillary start to grow in the retina. neovascularization is called this. often those delicate bleeding current vessels into the pigment particles. you might see some gloomy gnats, when they bleed a little. when it spills a lot, then all vision could be blocked. finally scar tissue, eventually aroused by the development of new capillary, can induce the retina to divide from either the rear in your eye. unless the new capillary interacts with ordinary fluid flow out from the eye, stress in the eye ball will accumulate. this can disrupt the nerves that brings stimuli (optic nerve) of your eye to your brain, contributing to macular degeneration.the body's effort to save its retina is proliferative retinopathy, but it can often lead to retina scarring and can cause the retina to detach itself, leading to blindness. modern eye care can help prevent blindness from arising as a result of proliferative retinopathy. b. stages of diabetic retinopathy stages descripti on image mild nonproliferative retinopathy microaneurysms occur at this stage. these are small pockets of globular swelling in the relatively small capillary of the eye. mailto:vijayashree.j@vit.ac.in knowledge based expert system for predicting diabetic retinopathy using machine learning algorithms retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication moderate nonproliferative retinopathy this is the phase where blocking of capillary occur. severe nonproliferative retinopathy the capillary which helps for the nourishment of eye are blocked thus signalling the retina to grow new capillary. proliferative retinopathy fresh capillary were proliferating, expanding within the retina, and into the vitreous gel. c. symptoms the initial stages of diabetic retinopathy will occur without signs or discomfort like many conditions of this nature. there will not be any direct effect on the vision when sickness progresses. macular oedema might led to maculopathy and affects vision if leakage induces swelling of the macular fluid. signs become obvious when the disease progresses, the normal retinopathy side effects to be observed also include:  blurry vision  spots and floaters in eye  double vision  eye pain d. risk factors someone living with diabetes may develop diabetic retinopathy. this may increase the probability of cultivating the eye condition:  high blood sugar level  high blood pressure  higher levels of protein content in urine  raised obesity in the blood  high levels of cholesterol  use of tobacco whoever has diabetes may grow diabetic retinopathy and other diabetes problems. the more a patient seems to get diabetes, the higher the risk of developing diabetic retinopathy. in addition, patients should always be informed that a rapid increase in blood glucose levels will result to retinopathy which is worse. in this scenario, a massive increase in blood sugar levels characterized by either a mmol / mol or percent reduction in hba c. e. complications of dr diabetic retinopathy includes the development of irregular blood vessels within that retina. abnormalities can cause major problems regarding vision::  vitreous!haemorrhage. the fresh capillary can bleed into the fresh, creamy-like stuff, covering the middle of your eye. where the rate of leakage is low, you can only see some dark spots (floaters). blood will fills the vitreous cavity in more severe cases and effectively block your vision. if the vitreous humor shrinks, these capillary can be weakened, causing them to bleed, which can contribute to the appearance of cobwebs in your eyes and make it harder to see. blood from a vitreous haemorrhage can dissipate, but any complications would require medical attention. vitreous haemorrhage on its own does not usually cause irreversible loss of vision. often the blood clears within a few weeks or months from the eye. without harm to your retina, the vision can revert to its former clarity.  retinal!tightening. the enlarged capillary correlated with macular degeneration facilitates the development of scar tissue which may remove the retina from the back of the eye.  glaucoma. new capillary must develop at the front of of your eye and collide to your eye's ordinary fluid flow, allowing pressure to build up in your eye (glaucoma). the above pressure may disrupt the nerve which carries images of your eye in your nervous system (antenna nerve).  blindness. diabetic retinal detachments, cataracts or both ultimately result in total vision lost. treatment may include one or more of:  laser therapy – to help new capillary rising  anti-vegf treatments – prevents the growth of new capillary but is a more expensive treatment. f. prevention diabetic retinopathy is not always preventable. regular eye tests, handling stable blood glucose and heart rate, and early intervention of vision problems may help prevent serious loss of vision, however. decrease the level your risk of diabetic retinopathy when you suffer from diabetes by doing the following:  monitor your diabetes: consider the daily routine part of healthy eating and physical activity. seek minimum minutes of medium aerobic exercise every week, for example walking. take drugs or insulin for oral diabetes as prescribed.  track the blood glucose level: you might need blood sugar monitoring and log them out many times a day— you may need more regular tests when you happen to be sick or under tension. request your physician how frequently you need to check sugar in blood.  keep levels of cholesterol control balance: healthy eating, doing daily workouts and dropping muscle mass will help. medication is also often required.  quit smoking: smoking will increase the risk of several diabetes problems.  beware of shifts in vision. whether you notice sudden changes in vision, or your vision is blurred, spotty or hazy, call your eye doctor right away. international journal of engineering and advanced technology (ijeat) issn: – , volume- issue- , april retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication g. statistics background retinopathy among diabetes sufferers some type of retinopathy is normal. a royal liverpool university hospital study examined type diabetes patients and , type diabetes patients and made the following results:  some form of retinopathy was found in percent of individuals with type one diabetes  some levels of the retinopathy was observed in . percent of individuals with type diabetes the greater occurrence of retinopathy in seen among individuals having type than those with type diabetes. table –ii: zone wise distribution of monitored diabetic patients, and area wise occurrence. at centers, diabetics found voluntarily assessed by citizens of society using a formal procedure that was given for evaluation by society. the findings were analyzed to assess the occurance of dr in the sample population and to classify age-related and historical risk factors such as length of diabetes, use of insulin, and other end-organ diseases using the chi-square method. a total of diabetics known to have been screened. in total, forms of data entry were deemed suitable for further review. approximately . percent were males, . percent were between the ages of and , nearly two-thirds of patients were from the western and southern zones, and more than half had diabetes over years. table iii. p occurance of diabetic retinopathy about diabetes period mellitus. table -iv occurance of diabetic retinopathy in patients with other end--organ disease. the effect of diabetic retinopathy on people who have other organ diseases is quite common. chart -i. zone-wise occurance of diabetic retinopathy; this represents the commonness of dr in patients in india. table -v. distribution of dr ii. literature review franklin et al.( ), this research work provides a technique for segmentation of retinal vessels that can be used in retinal image analysis of machines. this experimental technique can be a pre-screening tool to detect diabetic retinopathy early on. the methods used in analysis can be used anonymously in retinal images to classify and interpret vascular structures. fleming et al, , computerized image processing is extensively followed to shorten the task of grading images resultant from the diabetic retinopathy screening programs. attempting to correct exudates in retinal images is a primary goal for automatic identification being one of the indicators that the disease has advanced to a point that needs to be listened to as an ophthalmologist. diabetic images and normal images are taken from a fundus camera, processed and analyzed for back-propagation on a computer using a neural network. the network had been instructed in acknowledging functionalities of the retinal image. it evaluated the impacts of numerous network variables and automatic filtering strategies. knowledge based expert system for predicting diabetic retinopathy using machine learning algorithms retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication diabetic and normal images were then randomized to assess network performance. the model shows . % sensitivity and . % specificity .gardner et al. ) this paper attempted to detect exudates using neural network back propagation (karegowda et al., ). the publicly available diaretdb dataset for diabetic retinopathy was used in the assessment process. the optic disk is neglected to avoid the optic disk from interfering with detection of exudates. the model shows sensitivity of . %, specificity of % and accuracy of . %. dupas et al.( ) automated microaneurysm and exudate identification was functionalise to two small image databases storage that manually marked those lesions on. a computer-based diagnostic system was then developed and tested for dr and me risk recognition and rating, use of such a huge database comprising both ordinary and compulsive images and auto gradation comparison. preprocessing of comparison improvement is extended until four features are collected as input frequency variables, normal frequency differentiation, hue and several angle pixel resolution, to provide coarse segmentation as data variables utilizing fcm clustering tool (sopharak et al., ).. the model showed a sensitivity of . %, specificity of . % and accuracy of . %. this suggests an automated procedure for the identification of rough exudates, a lesion associated with diabetic retinopathy. using a statistical description, the algorithm based on their color and their edges, adding an edge identifying to address them. in this way, we test the method's robustness to make it suitable to a clinical setting. the model showed a sensitivity of . % (sanchez et al., ) rao et al.( ) among the difficult and important elements of managing primary open angle glaucoma (oag) is identifying glaucoma progression. it is caused by pressure accumulation inside the eye. detecting glaucomatous progression is important and demanding of handling firstly open angle glaucoma (oag). the model showed an accuracy of . %. an effective method for identifying exudates as hard and soft exudates in this article(rajput et al., ). to remove noise, the retinal photograph in the color space of cielab is processed. then after, the network of capillary is separated to allow the identification and removal of optic disks. using hough transform method, the optic disks are removed. the victim exudates are identified by using k- means clustering algorthim. the model showed a sensitivity of . % and accuracy of . %. classification schemes are developed and tested to deduct the presence or absence of dr. the detection rule is depends up on the problem of binary-hypothesis testing which clarifies yes / no decisions with the question. it also shows an overview of the bayes output optimality criterion applied to dr. on the real-world data, the proposed dss is evaluated. the model showed specificity of %(kahai et al., ) prasad et al. ( ) this analysis suggests that usage of morphological techniques and methods of segmentation to identify the capillary, microaneurysms and hard & soft exudates. the representation of the retinal fundus is split into sub frames. diverse attributes were derived through the image of the retinal fundus. on the extracted features hair wavelet transformations are applied. the main technique for the analysis of components is then applied for better selection of features. for the detection and classification of the images as diabetic or non-diabetic, neural network back propagation techniques were employed. the model showed an accuracy of . %. this paper explores and suggests a optimally modified morphological operators technique to be used on the low- contrast images of patients with diabetic retinopathy for exudate detection (sopharak et al., ). these automatically observed exudates are confirmed as compared with the hand-drawn ground-truths of professional ophthalmologists. the model showed a sensitivity and specificity is % and . %. kwasigroch et al.( ) to improve system performance, we suggested a separate class coding methodology that enabled details to be included on the value of the discrepancy here between expected performance and the targeted performance in the subjective function monitored during neural network training. we used normal precision measurements and a quadratic weighted kappa score to check classification capacity of the employed models. the model showed an accuracy of about %. we introduced a brief structure to the convolutionary neural network architecture by emerging a pre-processing layered and convolution layer for maximise the output for the convolutionary neural network classifier (khojaste et al., ). two image enhancement techniques such as contrast enhancement technique and adaptive histogram equalization with a minimal contrast technique. the model showed an accuracy of . %. priya et al.( )this paper proposed two models such as probabilistic neural network and support vector machine to diagnose diabetic retinopathy, and compares their efficiency. when diabetes progresses, a patient's vision can initialized to deteriorate and lead to diabetes retinopathy. two classes are established, namely nonproliferative diabetes retinopathy and proliferative diabetes retinopathy. pnn has an accuracy of . % and svm has an accuracy of %. the project's aim is to identify retinal micro-aneurysms and exudates using classifier for automated dr screening(gupta et al., ). it is necessary to implement an automated dr screening system for detecting dark lesions and bright lesions in photographs of digital funds. to detect retinal micro-aneurysms and exude images from retinal funds. the model showed a sensitivity and specificity of % and %, accuracy of %. geetharamani et al.( ) diabetic!retinopathy, a primary leads of blindness, is discussed in this study. for define diabetic retinopathy, a two-tier system is adopted. the suggested technique is evaluated by the uci machine learning repository on diabetic retinopathy drebechen dataset. the evolved rules are evaluated and the best rules are created through fold cross validation. diabetic!retinopathy is an eye disease caused by diabetes over the huge term (athira et al., ). in our paper we suggest an approach for the diagnosis of dr from r-cnn (regional convolution neural network) digital fundus images. r-cnn is highly accurate and resistant to lesion detection. throughout our work, we have implemented a new strategy, where the whole picture was segmented and only the regions of interest were taken for further analysis. international journal of engineering and advanced technology (ijeat) issn: – , volume- issue- , april retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication in our process we used layers of r-cnn, trained it on fundus images and checked images on them.both images were divided into two categories, i.e., with dr and without dr. this r-cnn (regional cnn) approach was found to be fast and accurate with an accuracy of approximately %. sopharak et al. ( ) a identifying of lesions in digital fundus images is required for development of an automated diabetic retinopathy screening system. microaneurysms are the first symptom of diabetic retinopathy in clinical treatment. microaneurysm numbers are used to denote the situation of the condition. early detection of microaneurysm may use to lower the risk of blindness. this study discusses a set of efficiently adapted morphological regulators which are used for microaneurysm detection on non-dilated pupils and significantly higher-contrast retinal images. the microaneurysms observed are checked when compared with the ophthalmologists ' hand-drawn ground-truth. as a outcome, . , . , . and . percent respectively were the sensitivity, specificity, precision and accuracy. diabetic retinopathy and blindness problems diabetic patients facing. as the number of patients with diabetes is steadily increasing, this also results in an increase in the data. therefore, the use of data mining methods is important to obtain the useful information and unseen knowledge. dm plays a major role in dr as it can be utilise to society's better health. for retinal fundus images there are many techniques and algorithms which help to diagnose. this paper discusses, classifies and compares the previously proposed algorithms and methods with a view to creating better and more effective algorithms. this paper presents a summary view of different data mining techniques which shows that knn and svm have given the best accuracies. this review paper can act as a resource for future researchers to use data mining techniques to predict diabetic retinopathy.(rathi, ) ananthapadmanaban et al.( ) the most frequent of eye disease is diabetic retinopathy, is affected by complications which occur when capillary in the retina weaken. when early detection is not achieved it results in vision loss. depending on the modeling objective many data mining techniques serve different purposes. the results of the various techniques for classifying data mining were tested using a different method. to predict early detection of diabetic retinopathy eye disease, we used naive bayes and support vector machine, and the results indicate that the naive bayes test was . percent accurate. brief insight into the identification of dr in human eyes using various forms of preprocessing & segmentation techniques is provided in this research article (kumaran et al.). the detection actually depends on the rnfl network region. if the total area of the nerve fiber is lower, it will be affected by diabetic retinopathy (dr) and if the region of the nerve network is larger, therefore diabetic retinopathy will not impact the eyes and is therefore normal. it is a well- known fact that diabetics play a critical role in the wellbeing of humans and affect all organs. one such organ that is in man's possession. the dr will lead to a loss of vision in the human eye as the optic nerve is connected to the brain. the retinal fundus images are widely used in disease-affected images to diagnose & interpret disease. raw images of the retinal fundus are difficult to process with machine learning algos. it is in this very context that a survey is being given here. kauppi et al.( ) the advancement of image processing techniques to a high level where the finding can be transferred from research laboratories to practice includes the following: protocols approved and applied to test the techniques, protocols similar to the strict medical care regulations, and medicine research. they suggested the first step towards a systematic review of methods for detecting diabetic retinopathy findings. diaretdb is at al complicated database in many respects but in reality it corresponds to the situation: the images are uncalibrated, the expert assessment is free form and the displays used to interpret the images are uncalibrated. diabetic retinopathy is among europe's most common causes of blindness. effective therapies do exist however. in more than per cent of all cases, correct and early diagnosis and proper treatment procedure will prevent blindness. as a screening tool for diabetic retinopathy, digital imaging is becoming available. in addition to providing a high-quality permanent retinal appearance record that can be used to track development or reaction to treatment and that can be checked by an ophthalmologist, digital images have the ability to be processed by predictive analytics systems. identified the primary creation of a method for providing automated of digital photograph taken as part of our clinic's daily monitoring of diabetic retinopathy. a deep-learning enhanced dr detection algorithm reaches significantly better performance than a earlierly reported, but virtually similar, technique not using deep learning (abramoff et al., ). deep learning algorithms have the suitability to develop dr screening performance and prevent this devastating disease from vision impairment and blindness. table vi: related work s/ n o. author year feature selection techniques used machine learning techniques used performan ce evaluation s. wilfred franklin , s. edward rajan ( ) segmentation technique back propagation algorithm accuracy: . % wong li yun( ) image processing techniques three-layer feedforwar d neural network % sensitivity, % specificity alan d fleming( ) multi-scale morphological process. retinopath y screening programme s % sensitivity . % specificity g gardner digital filtering techniques back propagation neural network. . % sensitivity, . % specificity asha gowda karegowda( ) decision tree and ga-cfs back propagation neural network . % sensitivity, %speci ficity, . % accuracy. b. dupas( ) embedded method (regression) grading of dr and risk of me . % sensitivity, . % specificity. knowledge based expert system for predicting diabetic retinopathy using machine learning algorithms retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication t. teng( ) pso image processing algorithms % sensitivity. gwenol´e quellec( ) correlation matrix cade algorithms group form akara sopharak( ) correlation matrix with heat maps fuzzy cmeans (fcm) clustering ppv . %,pl r . %, accuracy . % c. i. sánchez( ) univariate selection statistical classificatio n . % sensitivity p.v.rao( ) z score normalization artificial neural network (ann) . % accuracy dr. g. g. rajput( ) embedded methods k-means clustering technique. . % sensitivity, . % predictive value, . %. accuracy. p. kaha( ) wrapper technique decision support system (dss) % specificity deepthi k prasad( ) wavelet transformations back propagation neural network . % accuracy akara sopharak( ) wrapper method exudate detection and classificatio n %sensiti vity . %speci ficity manoj raju( ) detecting the laterality of fundus image deep learning application in classifying . % sensitivity , . % specificity , . % accuracy arkadiusz kwasigroch( ) embedded(regr ession) deep convolution al neural networks (cnn) % accuracy xiaogang l( ) filter methods convolutio nal neural networks (cnns) group form p. khojasteh embedded(las so regression) convolutio nal neural network architecture . % accuracy r.priya( ) svm probabilisti c neural network (pnn) . % accuracy kanika verma random forests technique density analysis and bounding box techniques. % accuracy swati gupta( ) recursive elimination computatio nal techniques % sensitivity, %specif icity %accura cy r. geetharaman i( ) wrapper and filter methods uci machine learning repository . % accuracy athira filter methods r-cnn (regional cnn) . % priya binary patterns svm . % chetoui local ternary pattern svm with a radial basis function kernel (svmrbf) , . % wan filter and wrapper techniques convolutio nary neural networks . % sopharak ltp bayes classifier . - sensitivity, . - specificity, . - precision . - accuracy rathi data mining techniques artificial neural network . % sopharak wrapper methods svm classifier . % rathi filter methods svm classifier % kandhasamy local binary patterns, colour moments svm classifier . ananthapadm anaban images by descriptors and hu moment of gist naive baye s and supp ort vector machine rapid miner tool . kumaran rnfl network region artificial neural network % akram npdr lesions gaussian mixture model . % kauppi pso image database, ground truth and evaluation methodolog y sensitivity % ege mahalanobis classifier bayes classifier, knn classifier % abramoff lesion detectors cnn sensitivity- . % specificity . % iii. methodology particle swarm optimization has been applied to numerous areas in optimization and in combination with other existing algorithms. this method performs the search of the optimal solution through agents, referred to as particles, whose trajectories are adjusted by a stochastic and a deterministic component. the selected features are then classified using the following international journal of engineering and advanced technology (ijeat) issn: – , volume- issue- , april retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication algorithms: decision tree, random forest, support vector machine. figure depicts the system architecture diagram fig. system architecture iv. result analysis the performance of the proposed work is analyzed using the following metrics: accuracy, sensitivity and specificity. table vii- comparison results of classifiers with regard to accuracy sensitivity in % classifiers no. of features ( ) no. of features ( ) no. of features ( ) svm . . random forest . . decision tree . fig. comparison results of classifiers with regard to accuracy table viii- comparison results of classifiers with regard to specificity specificity in % classifiers no. of features ( ) no. of features ( ) no. of features ( ) svm . . . random forest decision tree . . fig. comparison results of classifiers with regard to specificity table represents the comprasion between the classifiers terms of accuracy which is figured in fig. table ix-comparison results of classifiers with regard to sensitivity accuracy in % classifiers no. of features no. of features no. of features svm random forest decision tree fig. comparison results of classifiers with regard to sensitivity knowledge based expert system for predicting diabetic retinopathy using machine learning algorithms retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication v. conclusion thus the paper examines the diabetic retinopathy using pso feature selection algorithm on three different classifiers svm classifier accuracy ( ), sensitivity ( . ) and specificity ( . ). svm classifier have got the maximum metrics percentage for pso feature selection algorithm. references . franklin, s. wilfred, and s. edward rajan. "computerized screening of diabetic retinopathy employing blood vessel segmentation in retinal images." biocybernetics and biomedical engineering . ( ): - . . yun, wong li, et al. "identification of different stages of diabetic retinopathy using retinal optical images." information sciences . ( ): - . . fleming, alan d., et al. "automated detection of exudates for diabetic retinopathy screening." physics in medicine & biology . ( ): . . gardner, g. g., et al. "automatic detection of diabetic retinopathy using an artificial neural network: a screening tool." british journal of ophthalmology . ( ): - . . gowda karegowda, asha, et al. "exudates detection in retinal images using back propagation neural network." international journal of computer applications . ( ): - . . dupas, bénédicte, et al. "evaluation of automated fundus photograph analysis algorithms for detecting microaneurysms, haemorrhages and exudates, and of a computer-assisted diagnostic system for grading diabetic retinopathy." diabetes & metabolism . ( ): - . . teng, thomas, martin lefley, and d. claremont. "progress towards automated diabetic ocular screening: a review of image analysis and intelligent systems for diabetic retinopathy." medical and biological engineering and computing . ( ): - . . quellec, gwenolé, et al. "deep image mining for diabetic retinopathy screening." medical image analysis ( ): - . . sopharak, akara, bunyarit uyyanonvara, and sarah barman. "automatic exudate detection from non-dilated diabetic retinopathy retinal images using fuzzy c-means clustering." sensors . ( ): - . . sánchez, clara i., et al. "retinal image analysis to detect and quantify lesions associated with diabetic retinopathy." the th annual international conference of the ieee engineering in medicine and biology society. vol. . ieee, . . rao, p. v., r. gayathri, and r. sunitha. "a novel approach for design and analysis of diabetic retinopathy glaucoma detection using cup to disk ration and ann." procedia materials science ( ): - . . rajput, g. g., and preethi n. patil. "detection and classification of exudates using k-means clustering in color retinal images." fifth international conference on signal and image processing. ieee, . . kahai, pallavi, kameswara rao namuduri, and h. thompson. "a decision support framework for automated screening of diabetic retinopathy." international journal of biomedical imaging ( ). . prasad, deepthi k., l. vibha, and k. r. venugopal. "early detection of diabetic retinopathy from digital retinal fundus images." ieee recent advances in intelligent computational systems (raics). ieee, . . sopharak, akara, et al. "automatic detection of diabetic retinopathy exudates from non-dilated retinal images using mathematical morphology methods." computerized medical imaging and graphics . ( ): - . . raju, manoj, et al. "development of a deep learning algorithm for automatic diagnosis of diabetic retinopathy." medinfo. . . kwasigroch, arkadiusz, bartlomiej jarzembinski, and michal grochowski. "deep cnn based decision support system for detection and assessing the stage of diabetic retinopathy." international interdisciplinary phd workshop (iiphdw). ieee, . . li, xiaogang, et al. "convolutional neural networks based transfer learning for diabetic retinopathy fundus image classification." th international congress on image and signal processing, biomedical engineering and informatics (cisp-bmei). ieee, . . khojasteh, parham, et al. "introducing a novel layer in convolutional neural network for automatic identification of diabetic retinopathy." th annual international conference of the ieee engineering in medicine and biology society (embc). ieee, . . saha, rituparna, amrita roy chowdhury, and sreeparna banerjee. "diabetic retinopathy related lesions detection and classification using machine learning technology." international conference on artificial intelligence and soft computing. springer, cham, . . r.priya, p. aruna. ( ). svm and neural network based diagnosis of diabetic retinopathy. international journal of computer applications. volume – no. , march . verma, kanika, prakash deep, and a. g. ramakrishnan. "detection and classification of diabetic retinopathy using retinal images." annual ieee india conference. ieee, . . gupta, swati, and a. m. karandikar. "diagnosis of diabetic retinopathy using machine learning." journal of research and development . ( ): - . . athira, t. r., et al. "automatic detection of diabetic retinopathy using r-cnn." ( ). . priya, r., and p. aruna. "diagnosis of diabetic retinopathy using machine learning techniques." ictact journal on soft computing . ( ): - . . chetoui, mohamed, moulav a. akhloufi, and mustanha kardouchi. "diabetic retinopathy detection using machine learning and texture features." ieee canadian conference on electrical & computer engineering (ccece). ieee, . . wan, shaohua, yan liang, and yin zhang. "deep convolutional neural networks for diabetic retinopathy detection by image classification." computers & electrical engineering ( ): - . . sopharak, akara, bunyarit uyyanonvara, and sarah barman. "fine microaneurysm detection from non-dilated diabetic retinopathy retinal images using a hybrid approach." proc. of the world congress on engineering. vol. . . . rathi, p., and anurag sharma. "a review paper on prediction of diabetic retinopathy using data mining techniques." int. j. innov. res. technol . ( ): - . . sopharak, akara, et al. "automatic detection of diabetic retinopathy exudates from non-dilated retinal images using mathematical morphology methods." computerized medical imaging and graphics . ( ): - . . rathi, p., and anurag sharma. "a review paper on prediction of diabetic retinopathy using data mining techniques." int. j. innov. res. technol . ( ): - . . kandhasamy, j. pradeep, et al. "diagnosis of diabetic retinopathy using multi level set segmentation algorithm with feature extraction using svm with selective features." multimedia tools and applications ( ): - . [ ] ananthapadmanaban, k. r., and g. parthiban. "prediction of chances-diabetic retinopathy using data mining classification techniques." indian journal of science and technology . ( ): - . . kumaran, yogesh, and chandrashekar m. patil. "a brief review of the detection of diabetic retinopathy in human eyes using pre- processing & segmentation techniques." ( ). . akram, m. usman, et al. "detection and classification of retinal lesions for grading of diabetic retinopathy." computers in biology and medicine ( ): - . [ ] kauppi, tomi, et al. "diaretdb : evaluation database and methodology for diabetic retinopathy algorithms." machine vision and pattern recognition research group, lappeenranta university of technology, finland ( ): - . . ege, bernhard m., et al. "screening for diabetic retinopathy using computer based image analysis and statistical classification." computer methods and programs in biomedicine . ( ): - . . abràmoff, michael david, et al. "improved automated detection of diabetic retinopathy on a publicly available dataset through integration of deep learning." investigative ophthalmology & visual science . ( ): - . international journal of engineering and advanced technology (ijeat) issn: – , volume- issue- , april retrieval number: c / ©beiesp doi: . /ijeat.c . published by: blue eyes intelligence engineering & sciences publication authors profile j. jayashree received ug degree from anna university, tamilnadu and received pg degree from vit university, tamilnadu and phd from vit university. she is working as assistant professor senior at vit university, vellore, tamilnadu, india. her research interests include data mining, machine learning. she had published a good number of papers in reputed scopus indexed journals. sruthi r, i finished my schooling in sishya school, hosur. i am pursuing my b.tech computer science with information security in vit university vellore ponnamanda venkata sairam, i finished my schooling in st.ann’s high school. i am pursuing my b.tech computer science with information security in vit university vellore j.vijayashree received pg degree and phd from vit university,tamilnadu. she is working as assistant professor senior at vit university, vellore, tamilnadu, india. her research interests include data mining, machine learning. she had published a good number of papers in reputed scopus indexed journals. estj: an expert system for tourism in jordan procedia computer science ( ) – available online at www.sciencedirect.com - © published by elsevier b.v. this is an open access article under the cc by-nc-nd license (http://creativecommons.org/licenses/by-nc-nd/ . /). peer-review under responsibility of universal society for applied research doi: . /j.procs. . . sciencedirect international conference on communication, management and information technology (iccmit ) estj: an expert system for tourism in jordan nada s. husseina, musbah j. aqelb adepartment of computer science, faculty of informatiom technology, applied sceince private university, amman, jordan bdepartment of computer science, faculty of information technology, applied sceince private university, amman, jordan abstract an expert system for tourism in jordan (estj) was developed to recommend a suitable travel schedule that satisfies the tourist's interest. the system is useful for tourists, and tourism agencies to select the best package based on the proper time, budget, and preferences of required tourist places. the system was designed as a rule based expert system and implemented with a prolog language. the system was evaluated and tested with a specialist and the results concluded were up to the level of human expert. © the authors. published by elsevier b.v. peer-review under responsibility of universal society for applied research. keywords: expert system, artificial intelligent, tourism. . . introduction recently many organizations have shown a great interest in using information technology for smoothing their business and to compete with competitors [ ]. one of the major of information technology areas that finds a lot of interest is artificial intelligence. expert systems are one of artificial intelligence fields that use a specialized knowledge to achieve high performance decision in a particular area [ ]. expert systems found many applications in many fields, like, science, business, and medicine [ ]. tourism is the act of travel for predominantly recreational or leisure purpose and also refers to the provision of services in support of this act. however, it has also importance in terms of economic (tourism revenue), and governments are trying to attract tourists to the country by providing an appropriate atmosphere for tourists and tourism programs that attract people. the expert systems play a vital role in tourism industry. many expert systems were developed for supporting this area. for example, an expert system for tourist information © published by elsevier b.v. this is an open access article under the cc by-nc-nd license (http://creativecommons.org/licenses/by-nc-nd/ . /). peer-review under responsibility of universal society for applied research http://crossmark.crossref.org/dialog/?doi= . /j.procs. . . &domain=pdf http://crossmark.crossref.org/dialog/?doi= . /j.procs. . . &domain=pdf nada s. hussein and musbah j. aqel / procedia computer science ( ) – management was developed to recommend a suitable travel schedule that satisfies user input constraints such as time period, budget and preferences. the system can provide tourists with information on the route and the distance between any two towns in the region [ ].while moisuc diana-aderina, et. al., designed an expert system for rural tourism in maramures. the system was developed to evaluate the countryside hotels and ranking them. moreover, the user can evaluate the benefits of tourist destinations from several points of view and obtain information useful for decision making [ ]. however, yunus dogan, et. al., developed an expert system to support tourism sector in turkey, where tourists will be able to select the most suitable holiday places for themselves [ ]. however, sindhu b., et. al., designed an expert system for rating the ecotourism destination [ ]. in this research, an expert system will be developed to assist tourists and tourist agencies in jordan to select the best package according to the given budget, and preferred tourist places. khakzad developed an expert system that containing more than cities and rules for helping tourists to choose best destination town that has maximum matching with their important request [ ]. tourism industry at jordan plays significant role in its economy. since human expertise plays an important part in the activities of many sectors of the tourist industry. an expert system is developed to assist the system is useful for tourists, and tourism agencies to select the best package based on the proper time, budget, and preferences of required tourist places. . . tourism in jordan jordan attracted travellers since ancient times. today it is more charming and beautiful with its modern state, due to progress and prosperity. it allows visitors to enjoy their taste differently as they wanted. it is the lives of simplicity and unspoiled nature calm and charming area. jordan is one of the most important tourism countries in the world in general and in the middle east in particular because of its features and environment which makes it a country with a high tourist attraction [ ]. however, the tourism in jordan can be classified into the following categories: - historical - religion - beaches and water sports - reserves - entertainment - shopping - medical treatment. . . expert system design for tourism in jordan the proposed system, i.e., expert system for tourism at jordan (estj), is a rule based expert system which is implemented using prolog program. the system makes use of backward chaining for the inference engine and search in the knowledge base. the system has a graphical user interface where the user can interact with the system through this interface. the user is presented with a series of questionnaire in the window which the user has the option to answer in yes or no. the set of questions are prepared according to the type of tourist's trip to the required destination. now, according to the feedback given by the user, the search in the knowledge base for possible pattern matches. if there is a rule in the knowledge base which matches the given facts by the user, the system shows the possible diagnosis in the same window. the expert system was designed as a rule based expert system architecture and it is consisted from three parts as shown in figure ( ). these parts are as follows: . knowledge base . inference engine . user interface nada s. hussein and musbah j. aqel / procedia computer science ( ) – fig. : expert system architecture . knowledge acquisition the knowledge base can be considered as the heart of the expert system as all the required facts for building the rules are contained in the knowledge base. taking this knowledge as the source, rules for the expert system can be formed. the primary source for knowledge acquisition for the estj system was consultation with tourism's experts, internet, journals and tourism books. . knowledge representation the knowledge base contains facts and rules about tourism in jordan that acquired from specialists, tourism books, and journals issued from ministry of tourism. the knowledge in the knowledge base is represented as a set of rules. each rule specifies a relation, recommendation and has if part (condition) and then part (conclusion). whenever the condition part is true, then the conclusion part will be executed. these rules is then represented as a predicate logic and converted to a program as follows: rule representation if the tourist likes to have a short trip (one day) and likes to visit historical places and likes to visit romans civilization places then the recommended package starting from amman by bus jerash: an ancient city built by romans ajloun castel a nice historical place built by arabs the inference engine performs the reasoning process while the expert system finds a solution. it connects the rules given in the knowledge base with the facts provided in the knowledge base. the inference process is carried out in an interactive mode with the user providing input queries and responses to questions through the user interfaces that stored as dynamic information in the working memory. it tries to derive new information about the given situation using the rules in the knowledge and the situation specific knowledge in the working memory. . . expert system results and discussion the system provides a friendly user interface where the system starts its session by inquiring about the tourist programs by prompting the following questions during the consultation session: . trip duration (shot trip-one day) or multiple days. . trip type (e.g. historical, scuba diving, shopping, religion, etc.) . trip budget: minimum price, and maximum price for the total trip expenses user knowledge base inference engine facts expertise nada s. hussein and musbah j. aqel / procedia computer science ( ) – . trip destination: tourist sites that will be visited. . trip services: in case of multiple days trip (rank of hotel, single or double room, etc.) assume a tourist approached a tourist agency for a trip from amman–capital of jordan to a given destination like petra which is a historical site. there are two packages for this site. the short trip is one day trip directly to petra, while the other one is three days trip which includes visiting petra, wadi rum that located at the heart of desert where the tourists enjoy camping overnight, and aqaba city that has a beach and provides scuba diving entertainment. types of tourist’s agent will consult the expert system for the best package for this trip with minimum price. the system starts asking the following: do you plan to have one day trip, or long trip? (select: . one day trip . multiple days select: does the place you want to visit have historical places (y/n)? y which ancient civilizations do you like to see at jordan? (select: . romans . nabataean . arabian) do you like camping at desert (y/n) ? y do you like to visit religious places (y/n)? n do you like to visit places with beaches and that provide water sports (y/n) y do you like to visit places with shopping mall facilities and provides personal entertainment (y/n)? y do you like to stay in a hotel with one of the following ranks ? (select: . five star hotel . four star hotel . three star hotel) do you like to stay in a single or double room with your partner? (select: . single . double) the expert system accordingly will conclude the best package for the tourist with minimum cost as shown in figure ( ). the best choice for you is the following route: starting from amman by bus: . visiting petra for one night . camping one night at wadi rum . visiting aqaba city for one day . the cost of the trip is ($ ) which includes the following: . transportation . entry fees to the historical place – petra . camping one night at wadi rum . staying at five stars hotel with a breakfast meal. nada s. hussein and musbah j. aqel / procedia computer science ( ) – fig. : consultation session with expert system conclusion an expert system for tourism in jordan (estj) was developed using rule based expert system architecture. the system was designed to assist the tourist, and the tourist agencies to select the best package for the tourist according to the tourist's preferences. the system was evaluated by specialist and the results were very satisfying and up to the level of human expertise. references: [ ] razieh, ac., zahra, ac., "a review on expert systems and their usage in management", , vol. , issue , pp. - . [ ] samy, s., abu nasser, abu zaiter and ola, a., "an expert system for diagnosing eye disease using clips", journal of theoritical and applied information technology, , vol. , issue , pp. . [ ] borgohain, r., sanyal, s., "rule based expert system for diagnosis of neuromuscular disorders", cornell university library, . [ ] chauhan, r., "an expert system for tourist information management", international journal of computer science and communication, , vol. , no. . [ ]diana-aderina, m., simona-alina, s. and nela, n., "the use of expert systems in rural tourism in maramures", the annals of the university of oradea, , vol. , issue . [ ] dogan, y., and kut, a., "an expert system for summer tourism in turkey by using text mining and k-means++ clustering", ict innovations, . [ ] babu, s., subramoonium, s. and kv, k.., " design and development of an expert system for rating the ecotourism nada s. hussein and musbah j. aqel / procedia computer science ( ) – destinations", conference on tourism in india-challenges ahead, , pp. - . [ ] khazad h., " tourism expert system with clips using pfc", artificial intelligence and signal processing (aisp), th csa international symposuim on , may - , , ieee conference. [ ] badhad, i., "geography and tourist attractions", alwaraq publishing company, amman, jordan, . redalyc.a novel web-based human advisor fuzzy expert system journal of applied research and technology issn: - jart@aleph.cinstrum.unam.mx centro de ciencias aplicadas y desarrollo tecnológico méxico rafe, vahid; hassani goodarzi, mahdi a novel web-based human advisor fuzzy expert system journal of applied research and technology, vol. , núm. , febrero, , pp. - centro de ciencias aplicadas y desarrollo tecnológico distrito federal, méxico available in: http://www.redalyc.org/articulo.oa?id= how to cite complete issue more information about this article journal's homepage in redalyc.org scientific information system network of scientific journals from latin america, the caribbean, spain and portugal non-profit academic project, developed under the open access initiative http://www.redalyc.org/revista.oa?id= http://www.redalyc.org/articulo.oa?id= http://www.redalyc.org/comocitar.oa?id= http://www.redalyc.org/fasciculo.oa?id= &numero= http://www.redalyc.org/articulo.oa?id= http://www.redalyc.org/revista.oa?id= http://www.redalyc.org journal of applied research and technology a novel web-based human advisor fuzzy expert system vahid rafe* , mahdi hassani goodarzi , department of computer engineering, faculty of engineering, arak university, arak , - - , iran. department of computer engineering, islamic azad university- south tehran branch, iran. *v-rafe@araku.ac. abstract the applications of the internet-based technologies and the concepts of fuzzy expert systems (fes) have created new methods for sharing and distributing knowledge. however, there has been a general lack of investigation in the area of web-based fuzzy expert systems. in this paper, the issues associated with the design, development, and use of web-based applications from a standpoint of the benefits and challenges of development and utilization are investigated. the original theory and concepts in conventional fes are reviewed and a knowledge engineering framework for developing such systems is revised. for a human advisor to have a satisfying performance, expertise is a must. in addition, some of advisory rules are subject to change because of domain knowledge update. the human requests may have linguistic or crisp forms and a conventional expert system (es) is not able to overcome the fuzziness in the problem nature. in this research, a web-based fuzzy expert system for common human advisor (fes-cha) is developed and implemented to be used as a student advisor at the department's web portal. the system is implemented by using microsoft visual studio .net , mvc and microsoft sql server . keywords: fuzzy expert systems, web-application, common human advisor, total average. . introduction knowledge-based and decision making systems are the branches of artificial intelligence which are based on imitating the human demeanor in finding the pattern of solutions to problems. in the real world, if definite and straightforward solution cannot be found, human expertise is needed. experts often follow a trial-and-error approach for problem solving. since there is no specific solution for this kind of problems, defining a certain computer method for achieving the solution is difficult. therefore, expert systems are used to reach this goal. in these systems, the program consists of a set of rules. the knowledge in an expert human brain is also a set of if-then rules. m. h. goodarzi [ , , ] proposed the fuzzy application in student evaluating system, portfolio advisor system and educational advisor system. fuzzy concepts can convert multiple crisp inputs to specific linguistic variables and use fuzzy rules to infer. in [ ] the fuzzy-based advisor for elections and the creation of political communities was proposed. in [ ], a web-based fuzzy expert system is used to help inexperienced indian farmers in the use of pesticide for their farms. the initial version of this software was introduced in in a single-user form. in forums, usually a user starts a discussion and expresses his/her opinions and approaches to a particular problem. [ ] proposes a model for creating a fuzzy-based expert forum that intelligently responds to questions asked by users. finding the right broker at the right time is another issue that requires expertise. this may be the reason for which inexperienced investors loose in stock markets. in [ ] a stock expert system model is proposed. the goal of this system is to make a good suggestion based on information about goods and market in order to reduce the loss and increase the benefit. educational consulting system tries to mimic the behavior of the staff addressing the educational consulting issues. in [ ] a fuzzy expert system for intelligent tutoring systems with a cognitive mapping is proposed. human cognition has become one of the most attractive areas of research and application in artificial intelligence in which human susceptibility is emulated. in [ ] a new fuzzy method for hotel selection is introduced as a hotel advisory system.. in [ ] the student a novel web‐based human advisor fuzzy expert system, vahid rafe /  ‐ vol.  , february      achievements and education system performance in a developing country is proposed. the current paper, includes five major sections: in the next section some of related works are reviewed. the third section describes the fuzzy rule-based and decision making systems and introduces the proposed model. in section the proposed system is discussed in details. section includes a sample of the advisor system implemented in a university. finally, a conclusion is provided. . literature review in a real voting the total of both, positive and negative votes for the candidates are collected. a major problem for the voters is when they have to select their deputies from a large list of candidate. the problem is more serious in cases where the candidates are unknown to the voters. however, the creation of political societies interested in addressing political issues is a hurdle to overcome. in [ ] an advisor system for elections and creation of political communities based on fuzzy logic is proposed. in this approach the recommendation engine works with a modified fuzzy c-means algorithm and the sammon mapping technique used for visualization of recommendations. each year in india, many farms are destroyed due to pests attack and insufficient experience of the farmers. in , the loss was about . billion dollars. soybean pest expert system (soypest) [ ] is a fuzzy expert system that asks fuzzy questions in order to generate a web-based response for the user. soypest is created by using jess and gradually became more accurate by receiving feedback from the users and the experts. mutual information interchange and the creation ng of forums on the web are important issues which captivate many researchers. one of the best known content management systems (cms) tools for this purpose is vbulletin. there are reasons that support the possibility of receiving irrelevant answers, no answers at all, different confusing answers from several other users and unclear answers. these drawbacks may be considered as the achilles heel of such systems. in [ ], linguistic expressions are categorized and then n-gram algorithm is used to edit and convert the sentences to a proper format. this system supports languages and by default the questions are multiple choice questions. the strong point of this system is its gradual improving knowledge base, the extended number and the expanded fields of topics; however, no considerable effort is done to find the best answer and the problem is solved through partial simulation of the human brain. . fuzzy decision making system most humans face difficulties related to life rules and regulations during their problem solving process. these rules and regulations are changing every now and then therefore, an expert is needed to memorize these rules in order to be able to help humans in their issues. the state of each person regarding the rules and regulations may differ from that of other people. human state (hs) is a member of a fuzzy set with a degree of membership equal to hs . the first step: determining and fuzzificating the inputs to the system by using fuzzy rules. following are some examples of the fuzzy sets of the system on hand: fuzzy set for law in judgment system. fuzzy set for passed courses in university. fuzzy set for marks of selected courses in university. fuzzy set for the rank and grade of student in the entrance exam and so on… . one sample for fuzzification of the crisp variable of ta in university system total grade point average (gpa) of students can be categorized into these groups: a, b, c, d and e. this categorization can be expressed through linguistic terms as excellent, good, middle, weak, very weak. the second step: determining the degree of membership of linguistic terms including following phases: a novel web‐based human advisor fuzzy expert system, vahid rafe /  ‐ journal of applied research and technology phase . for each term, the value nearest to the numeric equivalent of the linguistic term which has the maximum degree of membership is selected. here, the highest value for the linguistic term “excellent” is and has the highest value in “very weak” fuzzy set. phase .for each term, the value (or values) which has (have) the membership degree of is (are) determined. phase . point with  are connected to points with  by lines to form a gaussian (exponential) membership function. in cases in which there is more than one point with  , a gaussian membership function is obtained. in this model, membership functions can be gaussian, z- shaped and s-shaped. the membership function of gpa variable is shown in figure . figure . student's gpa membership function for example: to fuzzify ta= into a linguistic variable, first of all, we write the formula for z- shaped and gaussian membership function.                               xc cxb ac cx bxa ac ax ax cbaxz ),,,( ( )         )( . expfgaussian  cx m ( ) then the result of equations for ta = are calculated giving the following equations: . ) / ( ) , , , ( )(           tax ac ax z  . ) exp( ) ( . exp )(        tax ta = with . )( tax exists in "very weak" fuzzy collection and with . )( tax is a member of "weak fuzzy" collection. with competitive method the ta is changed to weak linguistic variable. . rule extraction the core of the system is very flexible and can be applied in many advisory environments by substituting the knowledge-base of the system with an appropriate medical, judicial or sport, etc. advisory knowledge-base. for example, after an interactive negotiation with an advisor lecturer at a university, fuzzy rules can be elicited and used in the es. the previous section illustrates how gpa can be fuzzified. the following rules are the result of the negotiations: if the gpa is moderate and the number of semesters for which the student is registered is small and the courses that are not passed can be taken. and passed in one semester and the student has not given any pledges and the student has not received any disciplinary notices and the student has not reached the maximum time period for his/her studies and the student has marks between - and some other student has had conditions similar to this student a novel web‐based human advisor fuzzy expert system, vahid rafe /  ‐ vol.  , february      then to a large degree, it is possible that the student is allowed to continue his/her studies in the university by giving an official pledge of achieving a gpa over in the next semester. else according to the status of the student, the fuzzy system is not able to provide an answer. a human expert's opinion is needed. . publishing the system on the web the rapid increase of information on the internet is currently a key issue when one is looking for relevant information. the development of the world wide web and applying multimedia tools along accessibility of web sites from any place in the world makes feasible the design ofuser interface compatible with the web. many expert systems in different fields of expertise are developed (exsys corvid, soypest, etc.) however, few are applied. since linguistic terms and fuzzy sets are used, the process for inference should be done on the client rather than the server to reduce the server’s busy time. this procedure can be executed in browser by script languages like javascript, java, vb script, xml, ajax and applet. . one application of the proposed system in a university portal by implementing the proposed fuzzy advisor system in the university portal, before enrollment of the next semester, the students are checked and those who should be excluded are determined and prevented from registration for the next semester. this advisor system addresses his/her issue according to rules and regulations. in section some questions and answers, which were provided by the advisor and the student, are shown. . a look inside the system the proposed system is analyzed and designed by uml methodology and documents are generated with rational rose case tool. the software is built on -tier layers such that when one of the layers is reconfigured or rebuild, other layers don't change. the framework of system is shown in figure . figure . system framework. a) initially, the user selects the type of advisory service and enters crisp data in web application layer via a web browser. b) the input data related to the system is controlled for gpa, number of official notices received, educational level in university, criminal records (if any), type of illness in medical system (if any), etc. these are executed in business facade and business rules layer. c) a request for fuzzification the crisp variables and rules generation is submitted to knowledge-base of the system by data access layer with ado.net. then the linguistic variables are generated just by view select, stored procedure & user define function execution. this section makes a database abstraction and prevents sql-injection. d) linguistic variables are sent to inference engine and processed with mamdani model [ ]. this section is accomplished by one stored procedure in database, named ustpinference. e) fuzzy answers are defuzzified and crisp output values are generated. this step is accomplished by one-user defined function in database, named udfdefuzzifier. finally, the advisor system extracts answers to be shown to the user. data object model of system is shown in appendix . client side (javascript & jquery) server side (asp.net & mvc & c#) business layer (business façade & business rules) .net framework classes common (enumerations & data objects & data sets) data access layer (ado.net) database & databaseoobjects (stored procedures, user define functions, views) a novel web‐based human advisor fuzzy expert system, vahid rafe /  ‐ journal of applied research and technology figure . implementation architecture. . implementation environment the system is implemented in the following layers: web application layer: this layer includes web forms, web user controls, web component (infragestics grid control) and model view control (mvc). business facade, business rules layer: includes controls and business methods and attributes in c# classes:  fuzzy_decision.cs,  fuzzy_ruleinference.cs,  fuzzy_set.cs,  fuzzy_linguisticvariables.cs. data access layer: the class cls_dataaccess.cs supports ado.net and connectionless performance. first of all, users login to the portal site and select the advisor system link and select their problems category. system asks the questions related to the problems. collecting the answers provided by the user, the cha translates the user inputs to linguistic variables, makes the fuzzy rules and generate the fuzzy answer. then with segono model the fuzzy answer are defuzzified to crisp output and is reported to user. the implementation architecture of the system is shown in figure . in implementation architecture diagram sources can submit requests to the web server user, knowledge engineer and web service). then the web server controls the requests and sends the appropriate interface for this request. after entering user information, the fuzzy question generator creates the fuzzy questions for the user. user enters the inputs and submits the data to server. when the user inputs are received by the web server, with rule generator, the fuzzifier and knowledge base, the if- part for the rules is generated. then the inference engine refers to the system knowledge base and if a match is found for the pattern within a fuzzy statement by using the mamdani model, the fuzzy answers for thethen-part of the rules are generated. in case that the input data does not match any patterns in the rule base, an appropriate message stating that the system cannot find the answer is displayed. if the fuzzy answer is found, the system transfers that to the deffuzzifier and finally the crisp answer is reported to the user in the desired format including ms excel, html, xml, histogram and report file. this system can connect to another database for refining the output and generates the additional information. a novel web‐based human advisor fuzzy expert system, vahid rafe /  ‐ vol.  , february      . advantages of fuzzy expert systems the major advantage of these systems is that knowledge gradually turns into wisdom and can be used as a decision making tool in critical situations which replaces the conventional faq. some other features are:  more accessibility: many experiments can be done. simply an expert system is a mass production of experiments.  cost reduction: the cost of gaining experience by the user is decreased considerably.  risk reduction: the expert system can work in environments dangerous, harmful or unpleasant for human.  eternity: obviously, these systems don’t die.  multiple experts: an expert system can be the result of knowledge elicitation from several experts.  more reliability: these systems don’t get tired or sick, they do not go on a strike and they do not conspire against their managers. on the contrary, these are often done by human experts.  explanation capability: an expert system can explain the way in which the results are obtained. on the contrary, due to many reasons (fatigue, unwillingness, etc.) human experts are not able to provide such explanations all the time.  quick response: expert systems respond quickly.  responsibility in any condition: in critical conditions and/or emergencies an expert may be unable to make the right decision due to stress or other factors while an expert system’s decision making is not affected by these events.  experience base: an expert system can provide access to a massive amount of experience.  user training: an expert system can act like an intelligent tutor, i.e., problems are presented to the system and the way of reasoning can be obtained.  ease of knowledge transmission: one of the most important advantages of expert systems is its convenience to move the knowledge from the system to somewhere else on the globe. . one experimental result of common advisor system in university a student is going to be dismissed from the university and is going to lose a bachelor degree. the advisor asks a few questions to provide an answer. advisor: how many semesters have gotten a gpa under ? student: advisor: how many undertakings have you been given? student: advisor: how many disciplinary notices have you received from the university? student: advisor: how many semesters have you passed successfully? student: advisor: how many grades under have you gotten? student: advisor: how many course units have you passed out of ? student: advisor: what is your gpa? student: . the crisp student's answers, fuzzy values and the linguistic values of each question are shown in table . question answer fuzzy variable/(fv) )(x linguistic variable . high very low . low . high . very high . middle . . low table . fuzzification of the experimental input crisp variables a novel web‐based human advisor fuzzy expert system, vahid rafe /  ‐ journal of applied research and technology inference phase: if fv is high and fv is very low and fv is low and fv is high and fv is very high and fv is middle and fv is low then you are dismissed with a probability of percent. else not in knowledge base. . conclusion this paper glanced at the definitions and introductory concepts of fuzzy logic and fuzzy decision making and some implemented examples of such systems were presented. finally, a web- based student consulting expert system was proposed and its capability in enhancing the consulting process has been shown. acknowledgements this project supporting by the islamic azad university, south tehran branch references [ ] m. hassani goodarzi, "evaluating students' learning progress by using fuzzy expert systems", conf. rec. ieee th int. conf. international conference on natural computation (icnc' ) and the th international conference on fuzzy systems and knowledge discovery (fskd' ), tianjin, china , pp. - , aug . [ ] m. hassani goodarzi, " a web-based implementation of a portfolio advisor system based on fuzzy expert systems ", in conf. rec. ieee th int. conf. information & communication technology and system icts, surabaya, indonesia, pp. - , sep . [ ] m. hassani goodarzi, vahid rafe " educational advisor system implemented by web-based fuzzy expert systems", in journal of software engineerin and application(jsea), volume. no. , july . [ ] l. teran, " a fuzzy-based advisor for elections and the creation of political communities", in ieee journal - - - / , pp. - , . [ ] h. s. saini, r.kamal, a. n. sharma, "web based fuzzy expert system for integrated pest management in soybean", international journal of information technology, vol. , no. , pp. - , august . [ ] y. min huang, j. nan chen, y. hung kuo, y. -lin jeng, "an intelligent human-expert forum system based on fuzzy information retrieval technique", elsevier, expert systems with applications volume , pp. – , ( ). [ ] p. e. merloti, "a fuzzy expert system as a stock trading advisor", www.merlotti.com/enghome/computing/fes.pdf. [ ] m.h. fazel zarandi,m. khademian, b. minaei-bidgoli, "a fuzzy expert system architecture for intelligent tutoring systems: a cognitive mapping approach", scires, journal of intelligent learning systems and applications, , , pp. - doi: . /jilsa. . published online february . [ ] e.w.t ngai, f.k.t wat, "design and development of a fuzzy expert system for hotel selection", elsevier volume , issue , pp. - , may . [ ] j. h. marshall, ung chinna, ung ngo hok, "student achievement and education system performance in a developing country ", springer, educ asse eval acc, doi . /s - - -x, february . [ ] w. siler, j. j. buckley. fuzzy expert systems and fuzzy reasoning, john wiley & sons, inc(isbn - - - , ). a novel web‐based human advisor fuzzy expert system, vahid rafe /  ‐ vol.  , february      appendix : system object model .. . expert sysrems will applications ( ) - bivariate quality control using two-stage intelligent monitoring scheme crossmark - ibrahirn masood a.'!', adnan hassan 'faculty of mechanical and manufacturing engineering, universiti tun hussein onn malaysia, pant raja, batu pahat, johor, malaysia b ~ a c u l t y ofmechanical engineering, universiti teltnologi malaysia, utm sltudai, johor, malaysia a r t c l e i n f o a b s t r a c t article history: available online june -- keywords: balanced monitoring bivariate quality control statistical features synergistic artificial neural network two-stage monitoring in manufacturing industries, it is well known that process variation is a major source of poor quality products. as such, monitoring and diagnosis of variation is essential towards continuous quality improve- ment. this becomes more challenging when involving two correlated variables (bivariate), whereby selection of statistical process control (spc) scheme becomes more critical. nevertl~eless, the existing tra- ditional spc schemes for bivariate quality control (bqc) were mainly designed for rapid detection of unnatural variation with limited capability in avoiding false alarm, that is, imbalanced monitoring per- formance. another issue is the difficulty in identibing the source of unnatural variation, that is, lack of diagnosis, especially when dealing with small shifts. in this research, a scheme to address balanced mon- itoring and accurate diagnosis was investigated. design consideration involved extensive simulation experiments to select input representation based on raw data and statistical features, artificial neural net- work recognizer design based on synergistic model, and monitoring-diagnosis approach based on two- stage technique. the study focused on bivariate process for cross correlation function, p = . - . and mean shifts, p = k . - . standard deviations. the proposed two-stage intelligent monitoring scheme ( s-ims) gave superior performance, namely, average run length, arll= . - . (for out-of-control process), arlo = . - . (for in-control process) and recognition accuracy. ra = . - . %. this scheme was validated in manufacturing of audio video device component. this research has provided a new perspective in realizing balanced monitoring and accurate diagnosis in bqc. elsevier ltd. all rights reserved. . introduction in manufacturing industries, when quality feature of a product involves two correlated variables (bivariate), a n appropriate spc charting scheme is necessary to monitor and diagnose these variables jointly. specifically, process monitoring refers t o t h e iden- tification of process condition either in a statistically in-control or out-of-control, whereas process diagnosis refers t o t h e identifica- tion of t h e source variable(s) for out-of-control condition. in addressing this issue, the traditional spc charting schemes for bqc such as x (hotelling, ), multivariate cumulative sum (mcusum) (crosier, ), and multivariate exponentially weighted moving average (mewma) (lowly. woodall, champ, & rigdon, ; prabhu rungel-, ) are known t o be effective in monitoring aspect. unfortunately, they a r e merely unable to provide diagnosis information, which is greatly useful for a quality practi- tioner in finding t h e root cause error and solution for corrective action. since then, major researches have been focused on diagnosis * corresponding author. tel.: + . e-mail addresses: ibraliim@~~lli~n.edu.~ny (i. masood), atlnan@fl~m.ul~ii.~ny (a. hassan). urls: l~~~p://www.~~~i~~n~edu~my (i. masood), http://www.utm.my (a. hassan). aspect. shewhart-based control charts with bonferroni-type control limits (alt, ), principle component analysis (pca) (jacltson, ), multivariate profile charts (fuchs & ben,jamini, ), r decomposition (mason, tracy, & young, ) and minimax control chart (sepulveda & nachlas. ), among other, have been investi- gated for such purpose. further discussions on this issue can be found in lowry and montgomery ( ), i ). nevertheless, the existing traditional spc schemes were mainly designed by focusing on rapid detection of out-of- control condition (arl, * i ) but it has limited capability in avoid- ing false alarm (arl < ). fig. illustrates the concepts of imbalanced monitoring vs. balanced monitoring as the central theme for this investigation. based o n diagnosis viewpoint, an effective bivariate spc scheme should be able to identify the source variable(s) of out-of-control condition a s accurate as possible. nevertheless, it is difficult to cor- rectly recognize when dealing with small shifts ( ~ . standard deviation). chih and rollier ( ). chih and rollier ( ), zorriassatine, tannock, and o'brien ( ), chen and wang ( ) and yu and xi ( ). for examples, have reported less than % accuracy for diagnosing mean shifts at . standard deviation. among others, only guh ( ) and yu et al. ( ) reported the satisfied results ( > % accuracy). the imbalanced monitoring and lack of diagnosis capability as mentioned above need further investigation. in order to minimize erroneous decision making in bqc, it is essential to enhance the overall performance towards achieving balanced monitoring (rap- idly detect process variationlmean shifts with small false alarm as shown in fig. ) and accurate diagnosis (accurately identify the sources ofvariation/mean shifts). additionally, the bqcapplications are still relevant in today's manufacturing industries. in solving this issue, a two-stage intelligent monitoring scheme ( s-ims) was designed to deal with dynamic correlated data streams of bivariate process. this paper is organized as follows. section describes a modeling of bivariate process data streams and patterns. section presents the frameworlc and procedures of the s-ims. section discusses the performance of the proposed scheme in comparison to the traditional spc. section finally outlines some conclusions. . modeling of bivariate process data streams and patterns a large amount of bivariate samples is required for evaluating the performance of the s-ims. ideally, such samples should be tapped from real world. unfortunately, they are not economically available or too limited. as such, there is a need for modeling of synthetic samples based on lehman ( ) mathematical model. further discussion on data generator can be found in masood and hassan ( ). in bivariate process, two variables are being monitored jointly. let xl-i=(xl.i,. . .,xi- ) and x -i=(x . .. . . . x z . ~ ~ ) represent observation samples for process variable and process variable respectively. observation window for both variables start with sam- ples i = ( ,. . . - ). it is dynamically followed by (i + i ) , (i + ) and so on. when a process is in the state of statistically in-control, samples from both variables can be assumed as identically and indepen- dently distributed (i.i.d.) with zero mean ( p o = ) and unity standard deviation (go = ). depending on process situation, the bivariate samples can be in low correlation ( p = . - . ), moderate correla- tion ( p = . - . ) or high correlation ( p = . - . ). data correlation ( p ) shows a measure of degree of linear relationship between the twovariables. generally, this data relationship is difficult to be iden- tified using shewhart control chart as shown in fig. . on the other hand, it can be clearly indicated using scatter diagram. low corre- lated samples yield a circular pattern (circular distributed scatter plot), moderate correlated samples yield a perfect ellipse pattern, whereas high correlated samples yield a slim ellipse pattern. disturbance from assignable causes on the component variables (variable- only, variable- only, or both variables) is a major source of process variation. this occurrence could be identified by various causable patterns such as mean shifts (sudden shifts), trends, cyclic, systematic or mixture. in this research, investigation was focused on sudden shifts patterns (upward and downshift shifts) with positive correlation ( p > ). seven possible categories of bivariate patterns were considered in representing the bivariate process variation in mean shifts as follows: n ( ,o): both variables xi-i and x -i remain in-control. us ( ,o): shifted upwards, while x .i remains in-control. us ( . ): x .i shifted upwards, while remains in-control. us ( , l ) : both variables xl.i and x .i shifted upwards. ds ( ,o): shifted downwards, while x .i remains in-control. ds ( , l ) : x .i shifted downwards, while remains in-control. ds ( , l ) : both variables x .i and x .i shifted downwards. reference bivariate shift patterns based on mean shifts f . standard deviations are summarized in fig. . their structures are unique to indicate the changes in process mean shifts and data correlation. the degree of mean shifts can be identified when the center position shifted away from zero point ( ,o). . two-stage intelligent monitoring scheme as noted in section i , an integrated mspc-ann was combined in a single-stage monitoring scheme (direct monitoring-diagnosis) as proposed in chen and wang( ). niaki and abbasi ( ). and yli et al. ( ). the other schemes based on fully ann-based models as proposed in zorriassatine, tannocli, and o'bricn ( ), c u h ( ). yuand xi ( ) and el-midany et al. ( ) also can be classified as a single-stage monitoring scheme. in this research, two-stage mon- itoring scheme was investigated by integrating the powerful of mewma control chart and synergistic-ann model for improving the monitoring-diagnosis performance. framework and pseudo- code (algorithm) for the proposed scheme are summarized in figs. and respectively. it should be noted that an initial setting as fol- lows needs to be performed before it can be put into application: load the trained raw data-ann recognizer into the system. i. masood, a. hassan/expert systems with applications ( ) - y i s e n s ~ t ~ v ~ t v in mean s h ~ r detection capablhtv in false alann avo~dance i/ shorter arll represents fastel longer arlo repiesenis smaller detect~on of plocess mean sh~fts false alaim i t b u i current state i imbalanced n~onitoring: able to detect process mean shifts rapidly (arl, = ) but has limited capability to avoid false alarm (arlo ) i desired state (for this research) i / balanced monltonn. [reasonable fol cu~rent ~ r a c t l c e i able to detect process mean s h ~ f t s rapldly (arl, z, ) and nlalnta~n small false alarm (arlo >> ) % i i i ideal state i perfect balanced: able to detect process mean shifts as soon as possible (arli = ) i . . w~tllout tnggenng any false alarm (arlo = m) l ~ ~ fig. . current state and desired state towards balanced monitoring. set t h e values of means (p ,po ) and standard deviations (nol.ao ) of bivariate in-control process (for variables and x .i). these parameters can be obtained based on historical or preliminary samples. perform in-process quality control inspection until observa- tion samples (individual or subgroup) to begin the system. recognition window size is set to observation samples (for variables xl-i and x _i) since it provided sufficient training results and statistically acceptable to represent normal distribution. preli- minary experiments suggested that a smaller window size (< ) gave lower training result due to insufficient pattern properties, while a larger window size (> ) does not increase the training result but burden the ann training. rational to integrate the mewma control chart and the synergistic-ann model are based on preliminary experiments. generally, the mewma control chart is ltnown to be effective for detecting bivariate process mean shifts more rapidly compared to the x control chart. furthermore, it is very sensitive when deal- ing with small shifts (g . standard deviations). unfortunately, based on one point out-of-control detection technique, it gave lim- ited capability to avoid false alarm (arb ). this becomes more critical when the variables are highly correlated. in the related study, pattern recognition scheme using a synergistic- ann model gave better capability in avoiding false alarm (arl,, > ). as such, it can be concluded that process identifica- tion based on recognition of process data stream patterns (synergistic-ann model) is more effective compared to detection of one point out-of-control (mewma control chart). nevertheless, different techniques should have their respective advantages in terms of pointlpattern discrimination properties. in order to fur- ther improve the monitoring performance (arll =. l , arlo >> ), it is useful to combine both discrimination properties (mewma control chart and synergistic-ann recognizer) by approaching i. masood, r hassanlexpert systems with applications ( ) - shewhart control chart scatter diagram -- low correlation low correlation i y - i ' i' zo m ki ' $ inn nvrnbor orsamplcs - - - number uisamplcs xi moderate correction moderatc correlation i ..' zo so ~ ro l w - - - - nlmbcl oisdnlplcs numher urnampler x high correlation high correlation i i i i ~ ' imi . - - - nunlbel ofratnplcs number ulsamples x fig. . shewhart control charts and its respective scatter diagrams. two-stage monitoring and diagnosis. in the first stage monitoring, the mewma control chart is used for triggering bivariate process mean shifts based on 'one point out-of-control' as per usual. once the shift is triggered, the synergistic-ann recognizer will perform second stage monitoring and diagnosis through recognition of pro- cess data stream patterns that contain one of several out-of-control points. this approach is suited for 'recognition only when neces- sary' concept, that is, it is unnecessary to perform recognition while the process lies within a statistically in-control state. besides, recognition is only necessary for identifying patterns sus- pected to a statistically out-of-control state. besides producing smaller false alarm, this approach will also reduce computational efforts and time consumes for pattern recognition operation. . . m e w m a control chart the mewma control chart developed by lowry ct al. ( ) is a logical extension of the univariate ewma control chart. in the bivariate case, the mewma statistics can be defined as follows: [u:(ewmai, - ) ' + u j ( e w m a z i - p j - u : , ( e w m a l ; - p , ) ( e w m a z i - @ , j ] n m e w m a , = (u:.: - u:,, covariance matrix of mewma: h m e w m a the standardized samples (zli, zzi) with cross correlation func- tion ( p ) were used. thus, a = a = ; = p. notations l and i rep- resent the constant parameter and the number of samples. the starting value of ewma (ewmao) was set as zero to represent the process target (/a,,). the mewma statistic samples will be out-of-control if it exceeded the control limit (h). in this research, three sets of design parameters (a, h: . , . ; . , . ; . , . ) as reported in prabhu and runger ( ) were investigated. . . synergistic-ann model pattern recognizer synergistic-ann model as shown in fig. ( was developed for ( ) pattern recognizer. it is a parallel combination between two i. masood, a. h a s s a n l e x p e r t s y s t e m s w i t h applications ( ) - . . . . . . .:. . -: - - . n - . - . . . . - . - . . . . xi (down-shift) x (down-shift) - fig. . s u m m a r y o f bivariare s h i f t patterns for p = . , . and . individual anns that are: ( i ) raw data-based ann, and (ii) statisti- recognizers can be combined using simple summation: , = cal features-ann as shown in fig. . x(ord-~,of-~), where i = ( ,. . . . ) are the number of outputs. final let or^ = (ord-i,. . . , ord. ) and of = (of.l,. . . , of. ) represent decision ( ,,,,,,) was determined based on the maximum value seven outputs from raw data-based ann and statistical features- from the c o m b i n e d ~ ~ u t ~ u t s : ann recognizers respectively. outputs from these individual osynergy = max(o/l, . . . , ) ( ) . masood, r hassanlexpert s y s t e m s w i t h applications ( ) - bivariate shift patte~nsforrnoderate data correlation (p = . ) .- partially developed shift fully developed shift - . - . . . . x l (up-shift) - . - . . . . x i (up-shift) - . - . . . . x i (nonnal) h - . - w z . . b . .- . - - -. - - - . - . - . - . . . . - . - . . . . x (up-shift) xi (up-shift) xi (normal) u x i (down-shift) krfzfit x i (down-shift) fig. ( c o n t i n u e d ) multilayer perceptrons (mlp) model trained with back-propa- neurons, while statistical features input representation gation (bpn) algorithm was applied for the individual anns. this requires only neurons. the output layer contains seven neu- model comprises an input layer, one or more hidden layer(s) and rons, which was determined according to the number of pattern an output layer. the size of input representation determines the categories. based on preliminary experiments, one hidden layer number of input neurons. raw data input representation requires with neurons and neurons were selected for raw i. mflsood, a. hossanfexpert s y s t e m s w i t h applications ( ) - x i (down-sh~ft) xi ( down-shift ) - --- - l --- -- ! fig. ( c o n t i n u e d ) data-based ann and statistical features-ann. the experiments did not improve the training results but provided poorer results. revealed that initially, the training results improved in-line with these excess neurons could burden the network computationally, the increment in the number of neurons. once the neurons reduces the network generalization capability and increases the exceeded the required numbers, further increment of the neurons training time. i. masood. r hassanlexpert systems with applications ( ) - i......... ............................. ......................... ! first stage mewma , monitol-ing / control chart next yes (out-of-control) troubleshooting i and i i renew setting / / i ................................................ : i; t i ycs (out-of-control) i i ; identify the sources of mean shift i ................................................................................................................................... fig. . frameworlc for the s-ims. . . lnput representation lnput representation is a technique to represent input signal into ann for achieving effective recognition. there are various approaches could be used to represent input signal. raw data (standardized samples) is the basic approach (zorriassatine, tannock, & o'bricn. ). besides raw data, feature-based approach that involves extracted features from raw data is one of the successful technique in image processing (br~~nzcll & eriltsson, : i . ~ raw data- own.. kaw a a b a s e d statistical f e a t u r e s features- of.s i i fig. . synergistic-ann model. controlling or monitoring. insufficient denoising will distort recognizer for improving pattern discrimination capability. raw waveforms and introduce errors. inversely, excessive denoising data input representation consists of data, i.e., 'consecutive will over-smooth the sharp features of underlying signals by standardized samples of bivariate process (z .p l z .p ,. . . .z .p . recognizing them as noise or outliers. z .p ). statistical features input representation consists of last in this research, raw data and improved set of statistical features value of exponentially weighted moving average (lewma]) with were applied separately into training of the synergistic-ann a = [ . , . , . , . ], mean ( p ) , multiplication of mean with i. masood, a hassanlexpert systems w i t h applicatioiu ( ) - fig. . individual ann recognizer. w l g r e r - l w e r l w ( raw data-ann standard deviation (msd), and multiplication of mean with mean square value (mmsv). each bivariate pattern was represented by data as follows: lewma . -~~, lewm&. -p , lewm&. .pl, lewmao.io-pi, , & i , msdpi. mmsvpi, lewmao. -~ , lewmao. o-p . lewmao.is-p~. lewmao.io-pz~ ppz, msdpz, mmsvp . lewman features were talten based on observation win- dow = . the ewma-statistics as derived using eq. ( ) incorpo- rates historical data in a form of weighted average of all past and current observation samples (lucas cf saccucci. ): w l w - l w ~ l w ( statistical feature-ann xi represents the original samples. in this study, the standardized samples (zi) were used instead of xi so that eq. ( ) becomes: where < a ,< is a constant parameter and i = [ i , , . . . , are the number of samples. the starting value of ewma (ewmb) was set as zero to represent the process target (po). four value of constant parameter (a = . , . , . , . ) were selected based on a range within [ . , . ] recommended by lucas and saccurci ( ). besides resulting longer arlo, these parameters could influence the performance of ewma control chart in detecting process mean shifts. preliminary experiments suggested that the ewma with small constant parameter ( a = . ) were more sensitive in identify- ing small shifts ( ~ . standard deviations), while the ewma with large constant parameter (a = . ) were more sensitive in identify- ing large shifts (> . standard deviations). the msd and mmsv features were used to magnify the magnitude of mean shifts ( p ~ ~ p z ) : where (,u ,p ), ( a l , ) (,uf,p:) are the means, standard deviations and mean square value respectively. the mathematical expressions of mean and standard deviation are widely available in textbook on spc. the mean square value feature can be derived as in hassail ct al. ( ). further discussion on selection of statistical features can be found in masood ancl hassan ( ). . . recognizer training and testing partially developed shift patterns and dynamic patterns were applied into the ann training and testing respectively since these approaches have been proven effective to suit for on-line process situation (gi~h, ). detail parameters for the training patterns are summarized in tablcs ancl . in order to achieve the best training result for overall pattern categories, the amount of training patterns were set as follows: (i) bivariate normal patterns = [i x (total combination of data correlation)] and (ii) bivariate shift patterns = x (total combi- nation of mean shifts) x (total combinations of data correlation)]. in order to improve discrimination capability between normal and shift patterns, a huge amount of n ( ,o) patterns was applied into ann training. the us ( , l ) and ds ( , l ) pattern categories also require a huge amount of training patterns since it contain a more complex combination of mean shifts compared to the other bivar- iate shifts pattern categories. guh ( ) reported that the utilization of partially developed shift patterns in ann training could provide the shorter arll results. in order to achieve the best arll results for this scheme, different percentage of partially developed shift patterns were utilized for different range of mean shifts as shown in table . the starting points of sudden shifts (ss) were determined empiri- cally. the actual value of data correlation is dependent to the var- iability in the bivariate samples. the simulated values ( p = . , . , . , . . . ) as shown in table could only be achieved when the process data streams are in fully normal pattern or in fully devel- oped shift pattern. input representations were normalized to a compact range between [- ,+ ]. the maximum and the minimum values for normalization were talcen from the overall data of train- ing patterns. based on bpn algorithm. 'gradient decent with momentum and adaptive learning rate' (traingdx) was used for training the mlp model. the other training parameters setting were learning rate ( . ) learning rate increment ( . ). maximum number of epochs ( ) and error goal ( . ), whereas the network performance was based on mean square error (mse). hyperbolic tangent func- tion was used for hidden layer, while sigmoid function was used for an output layer. the training session was stopped either when the number of training epochs was met or the required mse has been reached. . performance results and discussion the monitoring and diagnosis performances of s-ims were evaluated based on average run lengths (arlo,arll) and recogni- tion accuracy (ra) as summarized in table . the arls results were also compared to the traditional multivariate statistical process control (mspc) charting schemes such as x (hotelling, ). mcusum (pignatiello & kunger, lf) ), and mewma (lowry et dl., ), as reported in the literature. in order to achieve balanced monitoring and accurate diagnosis, the proposed s-ims should be able to achieve the target perfor- mances as follows: i. masood. a. hossan/expert systems with applications ( ) - table . . parameters for t h e training patterns. pattern category m e a n shift ( i n standard deviations) data correlation ( p ) amount of training patterns n ( , o ) x i : . . , . , . , . , . x = x : . us ( ,o) x l : . , . ,. . .. . x x = x : . , . ,. . .,o.oo us ( , l ) x : . , o . o o , . . ..o.oo l o o x x = xl: . , . ,. . .. . us ( . ) xl: . , . , . , . ,. .., . x x = . x : . , . , . , . . . . .. . ds ( . ) x l : - . , - . .. . ., - . l o o x x = x : . , o.oo,.. ..o.oo ds ( , ) x : . , . ,..., . loox x = x l : - . , - . ,. . . , - . s ( , ) xl: - . , - . , - . , - . ,. . .,- . x x = , x : - . , - . , - . , - . ,. . ..- . table parameters for the partially developed shift patterns. range of mean shifts ( i n standard deviations) a m o u n t of partially developed shift patterns starting point of sudden shift (ss) sample t h sample t h sample t h table summary of monitoring-diagnosis capabilities. traditional mspc s-ims effective in monitoring (to identify out-of-control signal) limited to avoid false alarm (arlo r ) unable to identify the sources of variation (mean shifts) comparable to the traditional mspc in monitoring aspect capable to maintain smaller false alarm (arlo >> ) high accuracy in identifying the sources of variation (mean shifts) (i) ari, >> to maintain small false alarm in monitoring bivariate in-control process. (ii) short arll (average arll . for shifts range k . - . standard deviations) to rapidly detect bivariate process mean shifts. (iii) high ra (average ra % for shifts range . - . stan- dard deviations) to accurately identify the sources of mean shifts. . . monitoring performance in monitoring aspect. the arb represents the average number of natural observation samples of in-control process before the first out-of-control process signal exist as a false alarm. in other word, the arlo measures how long a spc scheme could maintain an in- control process running without any false alarm. on the other hand, the arll represents the average number of unnatural obser- vation samples before it is truly identified as out-of-control process signal. in other word, the arll measures how fast a spc scheme could detect process mean shifts. further discussion on this mea- sure can be referred to montgoluery ( ). ideally, a spc scheme should provide arlo as long as possible in order to minimize cost for investigating the discrepancy and trou- bleshooting while the process still within control. meanwhile, it should provide arll as short as possible in order to minimize cost for reworlts or waste materials. since t h e false alarm cannot be eliminated, the arlo >> is considered as the de facto level for balanced monitoring. in this research, the arls results of s-ims were simulated based on correctly classified patterns. generally, it can be observed that the smaller the mean shifts, the longer the arll values. this trend support the conclusion that process mean shifts with smaller magnitudes would be more difficult to detect. specifically, the s-ims indicated rapid detection capability for large shifts (shifts = a . arl = . - . ) and moderate shifts (shifts = a , arll = . - . ) with short ranges of arl . it was also capable to deal with smaller shifts (shifts=ilo. . a], arl = . - . , . - . ). in comparison to the x charting scheme, detection capability as shown by s-ims was faster for small and moderate shifts (shifts = . - ). in comparison to the mcusum and the mewma, it was slightly comparable in rapid detection for large shifts (shifts = . , arl,: s-ims = . - . , mcusum = . , mewma= . ) and moderate shifts (shifts= . o. arl,: s-ims = . - . , mcusum = . ; mewma = . ). similar trend can also be found when dealing with smaller shifts (shifts = la, arli: s-ims = . - . , mcusum = . ; mewma = . ). meanwhile, based on the range of arlo results ( p = . , . , . ; arlo= . , . , . ), the s-ims was observed to be more effective in maintaining smaller false alarm compared to the traditional mspc (arb r ). it should be noted that the results for medium and high correlation processes have exceeded as shown in the shewhart control chart (nels , ; shewhart, ). overall, it can be concluded that the proposed scheme indicated balanced monitoring performance. . . diagnosis performance in diagnosis aspect, the ra measures how accurate is a spc scheme could identify the sources of mean shifts towards diagnos- ing the root cause error and conducting troubleshooting. generally, it can be observed that the smaller the mean shifts, the lower the ra results. this trend supports the conclusion that diagnosis infor- mation for small process mean shifts ( . standard deviations) - i. masood, a. hassanjexpert systems with applications ( ) - table . performance comparison between t h e s-ims and t h e traditional mspc. pattern category mean shifts average run lengths recognition accuracv s-ims x ucl = . mcusum k = . h = . mewma i = . h = . s-ims xi x arlo for p = . . . . . arlo for p = . ra for p = . , . . . n ( ,o) . . . . . . . ( . ) ( . ) ( . ) n a arl, for p = . , . , . us ( ,o) . . . . . , . . . . . . us ( , l ) . . . , . , . . . . . . us ( . ) . . . . . , . . . . , . ds ( ,o) - . . . , . , . . , . . . ds ( , l ) . - . . , . , . . . . . . ds ( . ) - . - . . . . . . . . . . . average . , . , . . , . , . us ( ,o) .oo . . . . , . - . . - . . , . , . us ( , l ) . .oo . , . , . . . . , . us ( , l ) .oo . . . . . . . , . , ds ( . ) - .oo . . , . , . . . . . . s ( , l ) . - . . , . , . . , . , . ds ( , ) - . - . . . . . . . . . . average . , . , . . . . . . us ( ,o) . . . , . , . . - . . - . . , . , . us ( , ) . . . , . , . . . . . . u s ( , l ) . . . , . . . . . . . ds ( ,o) - . . . , . , . . , . , . s ( , l ) . - . . , . , . . . . . . s ( , ) - . - . . . . . . . . . . average . , . , . . . . . . us ( . ) . . . , . , . . , . , . us ( . ) . . . . . , . . . . . . us ( , ) . . . , . , . . , . , ds ( ,o) - . . . , . . . . . . . . s ( , ) . - . . , . , . . . . . . ds( ,l) - . - . , . . . . . average . . . . . . , . . . us ( ,o) . . . , . , . . , . , . us ( . ) . . . , . , . . . . , . us ( . ) . . . , . , . . , . , ds ( ,o) - . . . . . . . . . . , . s ( , l ) . - . . . . . . . . . . . d s ( , l ) - . - . o . . . . . . . . . average . , . , . . . . . . us ( . ) . . . , . , . . , . . . u s ( . ) . . . , . . . . . . , . us ( , l ) . . . , . , . . . . , ds ( ,o) - . . . , . . . . , . . . s ( , l ) . - . . . . , . . , . , . ds( ,l) - . - . ° . . . . . . . . . average . , . . . . . . , . grand average +( . - . ) . , . , . . . . , . note: design parameters for mewma control c h a r t in s-ims ( = . , h = . ). would be more difficult to identify. specifically, the s-ims indicated accurate diagnosis capability for large shifts (shifts = , ra = . - . %) and moderate shifts (shifts = , ra = . - . %) with high ranges of ra. although the results were slightly flange internal diametc degraded, it is still effective to deal with smaller shifts (shifts = [lo, . o], ra= [ . - . %, . - . %]). it should be ' noted that the ra results for medium and highly correlated pro- /' cesses were higher compared to low correlation process, which is i effective for practical case. since the traditional mspc charting i , n schemes were unable to provide diagnosis information, diagnosis \, " capability as shown by s-ims was absolutely capable in solving \~ such issue. overall, it can be concluded that the proposed scheme '> . indicated accurate diagnosis performance. table summarizes the comparison of monitoring-diagnosis capabilities between the groove & flange view roller head s-ims and the traditional mspc. fig. . functional features of roller head. i. masood, a. hassanjexpert systems with applications ( ) - extnision round bar tuining to rough size turning to size honing inner diameters n~ckel electroplating beal-ings assembly fig. . process plan for the manufacture of roller head. . industrial case study . broadly, the need for bqc could be found in manufacturing industries involved in the production of mating, rotational or mov- ing parts. investigation for this study was focused on the manufac- turing o f audio video device (avd) component, namely, roller head. this investigation was based on the author's working experience in manufacturing industry in johor, malaysia. in an avd, the roller head functions to guide and control the movement path of a film tape. inner diameters of roller head (id and id ) as shown in fig. a r e two dependent quality characteristics (bivariate) that need for joint monitoring-diagnosis. in current practice, such func- tional features are still widely monitored independently using shewhart control charts. it is unsure why the mspc was not imple- mented. based on the author's point of view, it could be due to lack of motivation, ltnowledge and sltills to adapt new technology. the process plan for the manufacture of roller head can be illus- trated in fig. . initially, an aluminium extrusion round bar was turned t o rough size (rough cut machining). then, it was turned to size (finish cut machining) to form functional features such as inner diameters, and groove and flange, among others. the machining of inner diameters was then extended into honing pro- cess to achieve tight tolerance for bearing assembly. hard coated surface was also necessary. as such, the machined work-piece was electroplated using nickel alloy before assembly. rtool rroller head <. ..-.---.-. ,./' tool bluntness loading error (decrement in id) (increment in id) fig. . process variation occurred in turning-to-size operation. automatically loaded into pneumatic chuck using a robotic system. bluntness in the cutting tool will cause gradual decrement in both inner diameters (id ,id ) with positive cross correlation ( p > ). in another situation. such inner diameters could be suddenly increased simultaneously and yields positive cross correlation ( p > ) due to loading error. based on two examples of bivariate process variation, industrial process samples were simulated into the s-ims for validating its applicability in real world. the first case study involves tool bluntness. the mean ( p ) and standard deviation (o) of bivariate bivariate process variation can be found in turning to size oper- in-control process were determined based on the first samples ation d u e to tool bluntness and loading error as illustrated in (observations lst- th). tool bluntness begins between observa- fig. . these disturbances will cause unnatural changes in t h e tion samples st- th. validation results are summarized in process data streams as shown in ' -able . the work piece is table ( , whereby the determination of process status (monitoring) table sources of variation in machining inner diameters. stable process process noise n ( ,o) tool bluntness ds ( , l ) loading error us ( . ) xi., ([dl) ~:: f l w w t . i-"_- -- n o r m a l _ n o r m a l " k m l-.. down-trend i -" down-trend h l up-shift scatter diagram i. - u xi ( down-trend ) i. masood r hassanjexpert systems with applications ( ) - - - table . - . inspection results based on tool bluntness case. i original samples standardized samples window range monitoring-diagnosis decision xi. (id xi- (id ) ((dl) zi- (id ) . . . . . . - . - . . . - . - . . . . . . . - . - . . . - . - . . . . . . . . - . . . - . - . . . - . - . . . - . . . . . . . . . . . . - . . . . . . . . - . - . . . - . - . . . . . . . . . . . . . . . - . - . . . . . . . . . . . . . - n . . - . - . - n . . . - . - n . . . . - n . . - . - . - n . . . . - n . . . . - n . . - . - . - n . . . . - n . . . . - n . . - . . - n . . . . - n . . . . - n . . - . - . - n . . . . - n . . . - . - n . . - . - . - n . . - . - . - n . . - . - . - n . . - . - . - n . . - . - . - ds ( i ) . . - . - . - ds ( i i ) . . - . - . - d ( ) . . - . - . - ds ( ) . . - . - . - ds ( ) . . - . - . - ds ( i i ) . . - . - . - ds ( i ) ( p l , p z ) = ( . , . ): (g,, ) = ( . x -~, . x note: observation samples highlighted in bold ( st- th) represent out-of-control process. table outputs of the scheme for tool bluntness case. reco~nition window (rw) - - - - - - - - - p decision based on mewma control chart rw p decision based on mewma control chart rw p n us (i ) us ( ) us ( ) ds ( ) ds ( ) ds ( ) note: bold value represents the maximum output of ann that determines pattern category. i. mosood. a. honanfexpert systems with applications ( ) - and sources of variation (diagnosis) are based on output of the scheme as shown in table . in t h e first samples, this scheme was able to correctly recog- nize t h e bivariate process data streams as in-control patterns (n). in this case, it was effective to identify bivariate in-control process without triggering any false alarm. bluntness of the cutting tool begins a t sample st, whereby this scheme was able to correctly recognize bivariate process data streams as down-shift patterns (ds ( , i ) ) starting from sample th (at window range st- th). in overall diagnosis aspect, this scheme was observed to be effective to identify the sources of variation in mean shifts without mistalte. the second case study involves loading error. similar as in the first case study, the mean ( p ) and the standard deviation (o) of bivariate in-control process were computed based on the first observation samples. loading error exist between samples th- th. validation results and related output of the scheme are sum- marized in tables s a n d respectively. based on t h e first samples, this scheme is effective to cor- rectly recognize the bivariate process data streams as in-control patterns (n). in this situation, the process was running smoothly without false alarm. improper condition of pneumatic chuck and robotic arm causes loading error between samples th- th. in this situation, this scheme was able to correctly recognize the bivariate process data streams as up-shift patterns (us ( , l ) ) start- ing from sample th (at window range: th- th). in overall diagnosis aspect, this scheme is capable to correctly identify the sources of variation in mean shifts without mistalte. table inspection results based on loading error case. i original sa~nples standardized samples window range monitoring-diagnosis decision xili ([dl x,.z (id z i ~ l (id ) zi-i (id . . - . - . . . - . . . . . . . . - . - . . . . . . . - . - . . . - . - . . . - . . . . . - . . . - . - . . . - . - . . . . . . . . . . . . . . . - . - . . . . . . . - . - . . . . . . . - . - . . . . . . . . . . . . . . . . - . . . . . - n . . . . - n . . - . - . - n . . . - . - n . . . . - n . . - . - . - n . . . . - n . . . . - n . . . . - n . . - . - . - n . . . . - n . . . . - n . . - . . - n . . . . - n . . - . - . - n . . . . - n . . . . - us ( ) . . . . - us ( ) . . . . - us ( ) . . . . - us ( ) . . . . - us ( ) . . . . - us ( ) . . . . - us ( ) . . . . - us ( ) . . . . - us( i ) . . . . - us ( ) . . . . - us( ) (/l,,/l~) = ( . , . ); ( u . u ) = ( . x - , . x note: observation samples highlighted in bold ( th- th) represent out-of-control process. i. masood, a. hassanlexpert systems with applications ( ) - table . - outputs of t h e scheme for loading error case. - - w i n d o w range (rw) - - - - - - - - - p . . . . . . . . . decision based o n mewma control c h a r t n n n n n n n n n rw - - - - - - - - - p . . . . . . . . . n n n n n n n n . . us ( ) . . us ( ) . . us ( i ) . . ds ( ) . . ds ( ) . . ds ( ) . . rw - - - - - - - - - p . . . . . . . . . n . . . . . . . . . us ( ) . . . . . . . . . us ( ) . . . . . . . . . us ( ) . . . . . . . . . ds ( ) . . . . . . . . . ds ( ) . . . . . . . . . ds ( ) . . . . . . . . . note: bold value represents t h e m a x i m u m o u t p u t of ann t h a t determines p a t t e r n category. g . conclusions this paper proposed two-stage monitoring approach in moni- toring and diagnosis of bivariate process variation in mean shifts. based on the frameworlc of s-ims that integrates the powerful of mewma control chart and synergistic-ann recognizer, it has resulted in a smaller false alarm ( a r b = . - . ), rapid shifts detection (arl = . - . ), and accurate diagnosis capa- bility (ra = . - . %) compared to the traditional spc charting schemes for bqc. since the monitoring and diagnosis performances were evaluated using modeling data, real industrial data were used for the purpose of validation. the case studies involved tool bluntness and loading error in machining operations, whereby the proposed scheme has shown a n effective monitoring capability in identifying the bivariate in-control process without any false alarm. the scheme also effective in diagnosis aspect, that is, in cor- rectly identifying the sources of mean shifts when process becomes out-of-control. based on the promising results, the s-ims could be a reference in realizing balanced monitoring and accurate diagnosis of bivariate process variation. in the future work, further investigation will be extended to other causable patterns such as trends and cyclic. aclcnowledgements the authors would like to thank universiti tun hussein onn malaysia (uthm). universiti teltnologi malaysia (utm), and ministry of higher education (mohe) of malaysia who sponsoring this work. references al-assdf, y. ( ). recognition of control c h a r t p a t t e r n s using ~nulti-resolution wavelets analysis and ncul-dl networlts. conlputers. a r ~ d i~~dustrialengir~ecr~~~g, , - . ail, f. b. ( ). multivariate quality contl-ol. in n. l. johnson s. ikotz (etls.). encvclopedia of.stotistica mcierices (vol. ). new yol-it: wiley. assaleh. i<.. ai-assal, y. ( ). features extraction a n d analysls for cldss~fying cdusable patterns in c o n t ~ - o l rhal-ts, coinputeis aitd industrial engirzrering, , .. . bersimis, s., psaraliis, . . i'anaretos, . ( ). multivariate ~ t a t i s l ~ c a l pi-ocess control c h a m : an ovel-view. @rulily and rellabilrty engineeririg i ~ ~ r e r r ~ a t i t r n a i , , - . bi-iinzell, i-]., eriltssnn. j. ( ). feature recluction for cla.ss~fication of r n u l t i d ~ m e n s ~ o n a l data. porter-ri recognition, ( , - - . clien, .. lu. s., lam, s. ( ). a hybrid system ibr si'c concurrenl patlel- recognilion. /ldvclnced enginc.eriilg inloi-~iialics. , - . clien. l. h.. & wang, t. y. ( ). ai-lificial neural networlts lo classify mcan shifts from niultivariate %%hart s ~ g n a l s . computers olld dush.iol engineerillg, , -~ . clieng, c. s., & cheng, - . '. ( ). i d e n t i f y ~ n g the source of variance shifts in the rnultivai-iafe pl-occss using neural netwol-its and s u p p o r l vector machines. expert ,y.stern.s wit applico ror .s. . , - . chih. w. h.. kollier, d. a. ( ). u ~ a g n o s i s charactcristics sor hivariale p a t t e r n recogliltion s c h c m e in spc. iruernalionul jouri~o! o j r uolity u r ~ d reliability molitr.yrnelit, l i ( ) . - . chih. w. i<., rollier. i). a. ( ). a ~ ~ i e t h o d o i o g y of pattern recogmtion s c h e m e s for t w o variables in spc. il~terr~ationol jou - a of q~ialin, and i(eliability mnnngenrent, ( ), - . crosiel-, r. b. ( ). m u l t ~ v a r ~ a t e g e n e ~ a l ~ z a t i o n s of c u ~ n u l a t i v e s u m q w l ~ t y c.ontrol schemes. 'ech ometrics. ( ), - . el-midany. i . t., el-baz, m. a,. & abtl-elwahecl. m. s. ( ). a proposccl irarneworlt for conlrol c h a r t p a t t c r n recognition in multivariale process using &i-tificial neural networlts. expert systems with aljplicutio~is, , - . fuchs. c.. benjamini, y. ( ). m u l t ~ v a r i a t e profile chart? for s t a t ~ s t i c a l ~ l r o c e s s control. technom~trics, ( ), - . gauri, s. i<., chaltraborty, . ( ). feature-l~ased recogiiltlon of control c h a r t patterns. conlputen und lndlrstriul erigir~eerin,y, . - . gauri. s. i<.. chakraborly, s. ( ). improved recognition of control c h a r t p a l l c r n s using artificial neural netwol-its. i ~ ~ t e n ~ n t i o r ~ c i l j o u n ~ u l ofadvnnted manlrfuucturirlg tei:hnolom,, , -- . guh, r. s. ( ). on-line identification and q u a n t i l i c a t ~ o n of mean shifts in b i v a r ~ a t e processes u s ~ n g a nn-bared approach. quality and reliabilrtj~ engineering i ~ ~ t e r n o t i o n a l . , -- . cull, r. s . ( ). simultaneous process mean and variance monitol-ing using artiiicial neul-al nctworlts. cornptrters rind ir~dlistrifll er~girirering, , - . guh. k, s., shiue, y. k. ( ). on-line idenlification of contl-ul c h a r t p a t t e r n s using selt:o~-ganizing approaches, lnternatioiiu! lour of production kesenrch, ( ). --- . h a c h ~ c h a , w., ghorbel. a. ( ). a survey of contl-ol-chart pattern-~'ecognitio~i i ~ t e r a t u r e ( - ) based oli a new conceptual clasrificat~on s c h e ~ n e . c(~nlputers o r ~ d incll~striol eli~@l~ee~-in,y, , - . hassan, a., nabi baksli, m. s., s h a h a r o u n , m. a,, k. jamaludin, h. ( ). impl-ovecl spc c h a r t paltern recognition using statislical features. li.i-norionoljouii~al o j producrion rescorch, ( ). - . .lotelling. - . h. ( ). multival-late quality contl'ol. in c. eisenhdrt, m. w . hastay, w. a. w a l l ~ s (eds.), teclli~iques ofstub.sticul ailnlysi.~. new york: mcgraw-i- . jacltson. j. e. ( ). a user's guide to principle compollerrt. nl'w jel-spy: wiley. iklosgen, w.. zytltow, j. m. ( ). haildbooi< of dotti minii~g arid knowledge discovery. lontlon: oxford university i'ress. i bv indicates that the key is i m p o r t a n t to establish that the patient has the disease under consideration. • a f < indicates that the absence o f the key is i m p o r t a n t to c o n t r a d i c t that the patient has the disease under consideration. confidence measure values are classified into four categories based on these two principles: i. c f > b v , af~>o. these keys are confirming. a confirming key in the patient profile contributes significantly to the likelihood score. its validation will lead to a high score. h o w e v e r even if the key is not validated, the disease m ay still figure p r o m i n e n t l y in the final differential diagnosis. . cf>~ bv, a f ( o . these keys are critical. critical keys have the most impact on determining the likelihood score. t h e validation o f a critical key contributes to a high score. t h e invalidation o f a critical key co n t ri b u t es negatively to the score by using the af. if a critical key is un k n o w n , a neutral position is taken. . cf ( b v . a f < o. these keys are contradicting. t h e validation o f a co n t rad i ct i n g key does not strongly confirm the existence o f the disease. t h e invalidation o f a c o n t r a d i c t i n g key can lead to a very low likelihood score. . c f < bv, a f i> o. these keys are minor. m i n o r keys are used for fine tuning the differential diagnosis and will play a greater role in the future screening process. this classification scheme a p p r o x i m a t e l y c o r r e s p o n d s to the following use o f eroking strength and frequency values in internist's scoring mechanism. • critical keys: eroking strength, frequency . • c o n t r a d i c t i n g keys: et,oking strength, l frequency . * confirming keys: et,oking strength, frequency . • m i n o r keys: et'oking strength, frequency i. each element in a slot list is evaluated accord i n g to the a b o v e classification. t h e m a t c h e r determines how well the patient profile fits the structure that is determined for this element. using the state o f the patient profile with respect to the attributes o f each element in this slot list and using the c f and a f factors, the m a t c h e r determines the cm o f this element. repeating this iterative process, all cm values o f the elements in a slot list are accumulated and normalized by the unique normalization_factor for the slot. t h e overall sum o f the slots determines the score o f a particular context and then the likelihood score o f the disease. t h e matching process for the slot values is illustrated below. in the code, context refers to the current name o f the context, hypothesis refers to the n am e o f the disease currently investigated l. . yal~inalp and l. sterling and patientslot has all the values that are currently k n o w n for the patient for a particular slot, such as symptoms. t h e first clause illustrates that the slots which are not applicable are n o t skipped over, with the assumption that they are completely satisfied for probability calculations. satisfy_slots(context, hypothesis, patient, patientslot, nolappl, . ). satisfy_slots(context, hypothesis,patient, patientslot, slot,slotprob) ,- slot' = = not_appl, satisfy_slot (context, hypothesis, patient, patientslot, slot,slot prob,o). satisfy_slot (context,hypothesis, patient, patientslot,[normalization_factor(n f)], slotprob,accprob) *- slotprob is accprob/nf. % normalize for a slot satisfy_slot (context, hypothesis, patient, patientslot, [ keyl key list], slot prob,acc prob) ,-- key = = normalization_factor(nf), satisfy_key(context, hypothesis, patient, patientslot, key,cm), accumulate (acc prob,c m,accprobnext), satisfy_slot (context,hypothesis, patient,patientslot, keylist, slotprob,accprobnext). satisfv__key determines whether a key is a single key o r a disjunction o f keys. t h e code for processing single keys is given below. th e find predicate extracts the values for a particular key from the patient profile. h a n d l e _ s i n g l e _ k e y ( c o n t e x t , h y p o t h e s i s , patient, p a t i e n t s l o t , k e y , keyvalues,c f,af, cm) ,--- i s_cl i r e c l k e y ( c o n text, key, keyval u es), find (key, patientvals, p a t i e n t s i o t ) , d i r e c t _ k e y ( c o n t e x t , h y p o t hesis, patient, patientvals, key, keyvalues, c f,af, c m). hand•e-sing•e-key( c•ntext'hyp•thesis•patient•patients••t•key'key•a•ues'•f•af'cm) ,-- n o t is_direct_key(context, key, keyvalues), extract_from (context, hypothesis, patient. patientslot, key, keyvalues, cf,af,cm). hand•e-sing•e-key( c•ntext•hyp•thesis•patient•patients••t•key•keyva•ues•cf•af•cm) .-- base_value(bv), not_known (context, hypothesis, patient, key, keyvalues, bv, c f,af, cm) direct_key( context, hypothesis, patient, patientvals, key, keyvalues, cf,af,cm). check_whether_absent(patientvals), absent_key(hypothesis,patient,patientvals, key,keyvalues, cf,af,cm), direclkey(context, hypothesis, patient, patientvals, key, keyvalues, cf,af,cm) .-- check_whether_present(patientvals), match_compare (context, hypothesis, patient, patientvals, key, keyvalues, cf,af,cm). . calculating confidence measures for keys this subsection describes how the individual c m s are calculated for individual keytuples. both direct keys and extractable keys are treated in detail. o u r description here is qualitative in nature. t h e exact formulae used can be found in ref. [ ]. the confidence measure o f a direct key is calculated t h r o u g h an extended c o m p a r i s o n o f the values in the key attribute list o f the d d with the patient values as shown below. t h e first stage is to calculate the patient sum. that is a score indicating how well the patient values match the attribute values. patient sums are only calculated for keys which actually a p p e a r in the patient profile. match_compare (context, hypothesis, patient, patientvals, key, keyvals, c f,af, c m ) ,-- compute_patient_sum (context, hypothesis, patient, patientvals,cf,af, key, keyvals,o,patientsum,contradiction flag), member (threshold (threshold),keyvals), find_normalization (keyvals,norm factor), compute_key_prob(contradiction flag, patientsum,norm factor,threshold,cf,af,cm). the m a t c h e r calculates patient sums as follows. first the terms used in the patient profile, which may be a mixture o f qualitative and quant i t at i v e terms such as days, are converted to the dijest d o m a i n d e p e n d e n t qualitative terms which are used in the d d s, fo r example short o r medium. t h e terms are then c o m p a r e d with the actual terms in the d d and exact matches an d c o n t r a d i c t i o n s are noted. t h e terms which exactly match are s u m m e d using weights which are given in the d d with respect to each attribute. this is illustrated with the co d e presented below. compute_patient-sum (context, hypothesis, patient, patientvals, cf,af, key, [threshold (t) ] ,totalsu m,totalsum,no). compute_patient_sum (context, hypothesis, patient, patientvals, cf,af, key, [ elementikeyvals] ,intersum,nextsum,contradiction flag) ,-- element'.,. = = threshold(t), match ( prevcontrad iction flag, context, hypothesis, patient, key, patientvals, element, intersum,totalsum), check_contradiction (prevcontradiction flag,context, contradiction flag,hypothesis, c f,af, key,totalsum,nextsum). match (no,context, hypothesis, patient, key, patientvals, element, i ntersu m,accsum) -- match-single_val(hypothesis, element, patientvals,attrcontr), accsum is intersum + attrcontr. match (yes, context, hypothesis, patient, key, patientvals, element, lntersum,accsum) , - is_a_contradiction (hypothesis, patient, key, patientvals), record_contradiction (context, hypothesis, patient, key). match_single_val ( h ypothesis,site (site, contr),aiivalues, contr) ,- member(site(patsite),aiivalues), appropriate_.site(hypothesis, patsite). match_single_val (anyconcept, parameter,aiivalues, contr) , - %generalized matching parameter =.. [name, parvai,contrl, findval = .. [name,someval], member(findvai,aiivalues), match_from_tables(anyconcept, name, parvai,someval). % sample facts appropriate_site(choledocholithiasis, righlupper_quadrant). appropriate_site (choledocholithiasis,epigestrium). match_from_tables(_,duration, days,short) *- number(days), days > = , days < . match_from_tables(_,duration, days,moderate) , - number(days), days > , days < . match_from_tables(_,duration, days,long) ,- number(days), days > . every direct key has a threshold, which is the m i n i m u m value o f the patient sum considered to adequately match the key. t h e second stage o f the m a t c h e r is to c o m p a r e the patient sum with the threshold set for this key. on the basis o f this c o m p a r i s o n , the m a t c h e r concludes whether the patient profile satisfies the a t t r i b u t e values completely, partially, o r c o n t r a d i c t s them, and calculates the c m accordingly. if the patient sum exceeds the threshold value, then we say that the direct key has been t'alidated. t h e cm value is this case is the c f value. f o r example, the at t ri b u t e values o f the patient in fig. indicates a sum o f points. this is equal to the threshold value for this key, t h erefo re the direct key pain is validated for this patient. t h e c m is then set to . . i f the patient sum is less than the threshold, and no c o n t r a d i c t i o n has been noted, the key has been partially z,alidated. t h e confidence measure is a normalized fraction o f the c f value. this is handled by compute-key_prob. m o r e details are in ref. [ ]. if a c o n t r a d i c t i o n has been noted, the value o f the cm differs depending w h et h er the absence factor o f the key is positive or negative. if the a f is negative, it is returned as the cm. otherwise the cm is the negative o f the c f value. this is handled by check_contradiction. l. i~. yal(~inalp and l. sterling we describe each o f the four categories o f extractable keys in turn, where the key appears in the patient profile: (i) special-purpose knowledge is used to handle the an at o m i cal o r physiological states that are indexed as a key, such as c o m m o n bile duct o b s t r u c t i o n as in fig. o r swelling o f the pancreas. some sample facts are given below. extract_from(context, hypothesis, patient, patientslot, key,keyvaluas, cf,af,cm) ,- anatomy(key), anatomy_test (context, hypothesis, patient, patientsiot, key, keyvalues,cf,af,cm). anatomy (key) ,- organ (key). anatomy(key) ,-- system (key, systemcomponents). anatomy (key) ,- system (sys, systemcomponents), part_of (key,systemcomponents). system (intra hepatic_ducts, [left_intra hepatic_duct,rig ht_intra hepatic_d uct] ). system (extra hepatic_ducts, [common_bile_duct,cystic_duct,pancreatic_duct] ). anatomy_test (context, hypothesis, patient, patientsiot,organ,present,cf,af,cf) ,- organ(organ), surgery(patient,surgerylist), not taken (surgerylist,organ). anatomy_test (context, hypothesis, patient, patientsiot,testcontext, (specification,facttodetermine),c f,af,cm) ,- test_illustrates (testcontext, specification, listoftests), prioritize(listoftests,finaltests), patient_satisfies (hypothesis, patient, patientsiot,testcontext, specification.facttodetermine, finaltests, cf,af,cm). f o r each state, the set o f relevant tests is determined along with their o rd er o f preference. t h e representation o f anatomical knowledge in d i j e s t has been designed to allow the m a t c h e r to find the necessary tests that would indicate the presence o f the specified state. f o r example, the m a t c h e r finds that ultrasound and c t tests are indicative for understanding the co n d i t i o n o f the c o m m o n bile duct when checking choledocholithiasis [ ]. after the necessar,v tests are found, it is determ i n ed whether the patient has taken the test. if he has not, the c m for this key is calculated using the c f and a f values, and varies depending in which o f the four categories the c f and a f values lie. i f the patient has taken the test, d o m a i n specific knowledge is used to determine whether the patient's test results satisfy the specified state. if so, the cm is set to the cf. otherwise, the c m is eq u at ed to a f because a conflict exists between the expected condition o f the patient and the patient profile. t h e r e is no possibility to partially validate these keys. f o r example, the results o f the u l t raso u n d for the patient in fig. are c o m p a r e d with the expected o u t c o m e s for the key c o m m o n bile duct. f o r the test results [ ], it is found that the c o m m o n bile duct o f the patient is very dilated. t h e r e f o r e , cm is eq u at ed to . . th e ultrasound also shows there are gallstones in the gallbladder. cm for this key is set to . . planning optimal order o f tests, prioriti-e, is a complicated issue, and could be the d o m a i n o f a n o t h e r expert system that would p e r f o r m in parallel to d i j e s t . currently, the tests are checked in sequential order. studies in decision analysis for developing clinical strategies similar to the one for the diagnosis o f extrahepatic obstructive jau n d i ce can be useful for the d ev el o p m en t o f this module. especially, the sensitivity, specificity, complications and the cost o f the individual tests have been investigated to devise different adaptive strategies for tests taking, represented as decision trees in ref. [ ]. we have used a~ailability as o u r criteria for ordering. (ii) keys referring to blood tests, such as amylase and bilirubin, are evaluated using possibility distribution curves which are graphs provided to us by o u r experts. first the m a t c h e r checks whether this is a key that requires curve fitting analysis by seeing whether a patient has taken the particular test. if not, the calculation o f the c m is carried out by considering the four classes o f c f and a f values as for the anatomical states. if the patient has the test, the patient value is checked by a disease specific possibility distribution curve, where each curve estimates the dijest likelihood that a patient with the particular test value has the disease being considered. the resulting possibility value is used along with the cf and af to determine the cm of this key. for example, if the patient's test result shows a particular positive possibility, this value is used to normalize the cf specified for this key. normalization is needed since the importance of this test result is specified with the cf, and how well the patient's result fits the expected value for the disease is determined by the curve. if the patient's test result contradicts the presence of the disease, invalidating the key, then the full af value is used as the cm value. curve fitting is actually not very suitable with prolog if speed and accuracy is required. it should be implemented as an external procedure. extract_from (context, hypothesis, patient, patientslot, key, keyvals, cf,af, cm) ,-- possibility_curve( key), curve_fitting (hypothesis, patient, patientsiot, key, keyvals,c f,af, cm). possibility_curve(key) ,-- blood_test(key). curve_fitting(hypothesis, patient, patientsiot, key, keyvals, ef,af, e m ) ,-- blood_test_analysis( hypothesis, patient, patientsiot, key, cf,af, cm ). blood_test_analysis(hypothesis, patient,patientsiot, bloodtest, cf,af, cm) ,-- (get_patient_val(bloodtest,serum,patientsiot, result); get_patient_val(bloodtest, blood,patientsiot, result)), blood_test (bloodtest, hypothesis, result, prob), calculate_cm (prob. hypothesis, patient, bloodtest, ef,af, e m ) . (iii) recall that compound keys refer to a collection of findings, for example prodrome. their analysis requires the matcher to consider each finding in the collection similar to the consideration of each attribute of a direct key. each finding for compound keys, though, has to be analyzed separately similar to an element of a slot. the collected result of all the findings determines the overall cm for this key. extract_from (context, hypothesis, patient, patientsiot, key, keyvalues, c f,a f,c m ) ,- c o n c e p l t a b l e ( c o n t e x t , key, keyeoncepts), satisfy_concept (context, hypothesis, patient, patientsiot, keyeoncepts, prob), concept_prob(prob,cf,af, cm). the sum of all the confidence measures of the findings that are related to this key is denoted cms. cms is tested with respect to an interval [ ,threshold) where the value of the threshold for compound keys is application-dependent. if cms lies within this interval, the presence of a finding can be neither validated nor invalidated, and is considered to be unknown. if the value is to the left of this region, the finding is invalidated and the overall cm is set to the af. otherwise, it is considered to be fully validated and the overall cm is set to the cf. this is handled by concept _prob. (iv) the rule names that are used within key tuples are evaluated by activating each rule, for example for li~'er tests and obstructit,e tests. these rules, which represent for example a group of tests, need to be evaluated considering domain specific dependencies of the tests. each rule is interpreted separately and the cm calculation varies for each. default behavior if the patient has not taken the test is similar to the default behavior for anatomical states and blood tests. for example, the rule obstructit'e tests in fig. is activated for the patient in fig. . the values of the tests of this patient is found to be sufficient for this rule. therefore, the cm for this key is set to the cf value, which is . . extract_from (context, hypothesis, patient, patientsiot, key, keyvalues, c f,af, c m ) ,-- call_proc ( [ key, context, hypothesis, patient, patientsiot, keyvalues, c f,a f,c m ] ). / * c a l l a n y procedure passed as p a r a m e t e r ' / call_proc([procnamellist]) , - proc = .. [procnamellist],proc. the matcher has a default behavior for evaluating keys which are not covered by the above discussion, for example a direct key in the dd which does not appear in the patient profile, or a compound key for which no information is known. the cms of these keys are determined with l. ci. yal(;inalp and l. sterling respect to the f o u r categories o f c f a n d a f values. t h e crucial categories o f critical keys a n d c o n t r a d i c t i n g keys are chosen so as not to c o n t r i b u t e to the overall sum. t h e confidence m e a s u r e c m is calculated as follows: n o l k n o w n (context, hypothesis, patient, gey, keyvals, bv,cf,af,cm) ,-- % minor keys cf < bv a f > = , cm is(cf + af)/ . not_known (context, hypothesis, patient, key, keyvals, bv, c f,a f, ) ,- % critical keys cf > = bv, a f < , record_question ( [ context, h ypothesis, patient, key, keyvals] ). not_known (context, hypothesis, patient, key, keyvals, bv,cf,af, ) ,- % contradicting keys a f < o , cf < bv record_possible_contra (context, hypothesis, patient, key). not_known( context, hypothesis,patient, key, keyvals, bv, cf,af,af) ,- % confirming keys a f > = c f > = bv, record_u nknown ( context, hypothesis, patient, key). f o r e x a m p l e , the exact location o f the o b s t r u c t i o n c a n not be d e t e r m i n e d by u l t r a s o u n d for the patient in fig. [ ]. this key is a critical key. t h e r e f o r e , the c m value is set to by the default values as described a b o v e . . . orerall likelihood score t h e o~erall likelihood score o f a slot ls~o,, as m e n t i o n e d earlier, is the sum o f the c m for each key a n d n o r m a l i z e d by the specific n o r m a l i z a t i o n f a c t o r o f the slot. t h e n o r m a l i z a t i o n factor, n f , is defined as follows ~ h e r e n is the n u m b e r o f elements in a slot. we a s s u m e that not all absence f a c t o r s are zero. " ,~af,, a f >/ , n f = e f = , -i.cf,, a f < . t h e weighted sum, ws,, can be defined as the best case where all the elements o f the slot i is validated. t h u s , ws, = e~'= t c f , . t h e r e f o r e , n f , ~< ws,. with this relation, the n o r m a l i z a t i o n helps to increase the c o n t r i b u t i o n o f slot i to the likelihood o f the overall context. t h e score might be g r e a t e r t h a n i with d a t a that c o n f i r m s all the expected values o f a slot in a disease descriptor. t h e overall likelihood o f a c o n t e x t is thus defined as n \ t ls o~, l c = j=l n v where n v equals the n u m b e r o f valid slots in a context. let us illustrate this calculat,on b~ using the e x a m p l e disease in fig. a n d the patient in fig. . i f the l a b _ t e s t s slot is considered, it is seen that the n o r m a l i z a t i o n _ f a c t o r . . is calculated as described above. t h e calculation o f c m s for each o f the keys in this slot is illustrated in section . . respective b , the~ are . , . , a n d . . t h e sum o f these c m s is . . using these values, lsk,,, is set to . . since there is only one slot in this context, l~b_, .... is equal to . . . . the patient anao,sis when the m a t c h e r calculates the likelihood scores o f a disease, special i n f o r m a t i o n related to the patient with respect to each disease is recorded a l o n g with the likelihood scores, t h i s d i j e s t information is used to produce an evaluation report a b o u t the status o f a patient. it consists o f the list o f findings which are expected but not present in the patient data, which are contradictory to the evaluated disease, and the important concepts which have not been validated during the analysis o f the m a t c h e r . the findings o f the evaluation are divided into four categories, questions, contradictions, possible contradictions and unknowns. to record this information, again the four categories o f c f and a f values are used. the code in the previous section is suitably adapted. the evaluation report can be used to guide the subsequent stages o f clinical diagnosis in the screening process shown in fig. . for example, the missing necessary tests to check a specific condition that have not been performed are suggested by questions for a disease. contradictions are the set o f facts in the patient profile that contradict the existence o f the disease. possible contradictions are the unknown classes o f information which might be critical. they can contradict the disease if their definite absence is proven. u n k n o w n s is the category o f data that can be used for confirmation but are u n k n o w n at the time o f evaluation, . p e r f o r m a n c e o f d i j e s t the development time for d i j e s t was a b o u t nine m o n t h s including our learning a b o u t aspects o f jaundice, the diseases and the related a n a t o m y and the physiology. the knowledge represen- tation scheme and the uncertainty reasoning mechanism reflect our perception o f medical concepts and clinical reasoning provided by our experts. d i j e s t has been tested with cases taken from medical text books and real patient records. for example, table i shows a differential diagnosis produced by dijest for a patient with choledocholithiasis. the medical history o f this patient is shown by fig. . during testing, the evaluation o f all the d o m a i n diseases were included. in clinical use, a threshold m a y be used to inhibit unlikely diseases. the analysis shows that choledocolithiasis is given the highest likelihood score by dijest, even though it does not get the highest score in each context. the score for acute cholecystitis shows the way large absence factors can prevent a disease from being considered seriously as explaining the jaundice. the scores from the contexts o f clinical examination and lab tests strongly suggest that cholecystitis could explain the jaundice, more so than choledocolithiasis, but the patient's history strongly contradicts the disease. the evaluation report o f this patient points out for example the lack o f information about critical findings o f hepatitis, such as the presence o f a prodrome, or the exposure to the use o f needles in the past. the evaluation report is not shown here. later on in the course o f the disease, the same patient contracted pancreatitis, directly caused by the choledocholithiasis. we added new test results to the patient profile and re-ran dijest. the result o f the second differential diagnosis is given in table . the only changed scores are o f those diseases related to the pancreas. note especially that the likelihood score o f pancreatitis has significantly increased. table demonstrates the ability o f dijest to cope with multiple diseases. knowledge is still necessary, for example, to realize that hepatitis and choledocholithiasis do not in general co-exist, whereas choledocholithiasis m a y cause pancreatitis. such reasoning, which would form part o f the screening process, allows us to place more significance on the score for pancreatitis than for hepatitis even though it is actually marginally lower. table l table likelihood scores for patient i disease history clinical choledocholithissis i . . viral hepatitis - . . hepatitis - . . acute cholecystitis - . . pancreatitis . . pancr, pseudo cyst . . cirrhosis . . pajncreatic cancer - . . tests total score . . . . . . . . . . . . - . . - . - . ii ii likelihood scores for patient disease history clinical tests total score choledocholithiasis . . . (). viral hepat, itis - . . . . pancreatitis . . . . hepatitis - . . . acute cholecystitis - . . . . cirrhosis . . - . . pancr, pseudo cyst . . - . - . pancreatic cancer - . . - . - . l. i~. yal~inalp and l. sterling d i j e s t was i m p l e m e n t e d by using p r o l o g c o n s t r u c t s which are s t a n d a r d in a l m o s t all prologs. it currently runs under sicstus a n d q u i n t u s prologs. in terms o f speed, p r o d u c i n g a table such as a b o v e a n d the e v a l u a t i o n r e p o r t takes only a few seconds on the average. . c o n c l u s i o n s t h e features o f d i j e s t in its c u r r e n t state can be s u m m a r i z e d as follows. medical k n o w l e d g e is represented declaratively. t h e d o m a i n specific knowledge a n d d o m a i n specific r e a s o n i n g is clearly distinguished f r o m d o m a i n i n d e p e n d e n t k n o w l e d g e by the m a t c h e r by using different types o f keys. t h e c o m p l e x medical k n o w l e d g e related to the diseases, the characteristics o f different testing p r o c e d u r e s a n d the basic a n a t o m i c a l a n d physiological structure o f the b o d y are all represented i n d e p e n d e n t l y o f p a t i e n t i n f o r m a t i o n a n d illustrate characteristics o f jaundice. using p r o l o g e n a b l e d us to reach o u r objective, to have this s e p a r a t i o n and write a specialized interpreter very easily. t h e interpreter has also been generalized to handle d o m a i n s o t h e r t h a n d i j e s t by c u s t o m i z i n g the general m a t c h i n g capabilities o f the interpreter. r e p r e s e n t i n g the likelihood estimates by using two s e p a r a t e factors, c o n t r i b u t i o n and absence factors, can distinguish between valid, invalid, u n k n o w n and a b s e n t data. d i j e s t presents very realistic likelihood estimates o f the presence o f the c a n d i d a t e diseases by e v a l u a t i n g the patient profiles, which m a y be incomplete. o f special i m p o r t a n c e is the calculation o f likelihood scores o f the individual c o n t e x t s a n d their effect on the final diagnosis. d i j e s t also e m p h a s i s e s significant f a c t o r s in the e v a l u a t i o n o f each disease. c o n t r a d i c t o r y findings a n d i m p o r t a n t d a t a which m a y be required for further e v a l u a t i o n o f the patient are noted. d i j e s t is very p r o m i s i n g in the early detection o f co-existing diseases in a patient a n d provides g o o d likelihood estimates in the cases with multiple diseases. t h e m o s t difficult task in d i j e s t is to o b t a i n the c o n t r i b u t i o n a n d absence factors for different keys. especially, representing the experts" qualitative view o f the subject by using those factors needs successive e x p e r i m e n t s a n d a d j u s t m e n t . a w e a k p o i n t o f d i j e s t is its neglect o f unexplained factors that are c o n t a i n e d in the patient profile. t h e presence o f a screening process for presenting the results o f m a t c h e r in a user-oriented m a n n e r a n d for r e m o v i n g r e d u n d a n t i n f o r m a t i o n would e n h a n c e the p e r f o r m a n c e o f d i j e s t . t h e consistency checking is also only partially complete. at this stage, however, d i j e s t is e n c o u r a g i n g in its expressive p o w e r for medical knowledge a n d by p r o v i d i n g useful likelihood estimates to indicate the presence o f d o m a i n diseases. it has p o t e n t i a l for detecting the co-existence o f multiple diseases. it is unique in b o t h its knowledge r e p r e s e n t a t i o n scheme a n d r e a s o n i n g with uncertainty. . cknowledgements--we would like to thank our medical expert, professor david ransohoff, for providing the medical knoaledge embodied m dijest. we are grateful for his ~aluable time in attending the knowledge engineering sessions and ackno~ledge his influence in shaping dijest. drs arnold shmerling and lawrence widman also supplied valuable medical insights. dr len samuels commented on an earher draft and pro,,ided the correspondence between contribution and absence factors of dijest and the scoring mechamsm of internist. we also thank yuval lirox for inviting us to submit this paper to the specml isssue of computers & mathematics with apphcations. r e f e r e n c e s s. pauker, a. gorry. j. kassier and w. schwartz, to,~ards the simulation of clinical cognition: taking a present illness by computer. m. d. med. , - ( ). . b g. buchanan and e. shortliffe. rule based expert systems the ycin ex'perlments o f the stanford programming prolect. addison-wesley, reading. mass. ( ). . r. miller, h. pople and j d. meyers, internist i, an experimental computer based diagnostic consultant for general mternal medicine. new engl. j, med. , - ( ). . f. e. mesarie, r. miller and j. d. meyers, intermst-i properties: representing common sense and good medical practice in a computerized medical knowledge base. computers biomed. res. , ~, ( ). . p. szolovitz and s. pauker, categorical and probabilistic reasoning in medical diagnosis. art!l intell. ii, - ( ), . harrtson's principles o f internal medicine, i ith edn. mcgraw-hill, net york ( ). . l i~i. yalqmalp. uncertainty reasomng in a medical expert system: dijest m.s. thesis, department of computer engineering and science, case v~estern reserve university, cle~,eland, ohio, ( ). . j. richter, m. silverstein and r. shapiro. suspected obstructive jaundice, a decision analysis of diagnostic strategies. ann. internal med. , - ( ). a semantic fuzzy expert system for a fuzzy balanced scorecard a semantic fuzzy expert system for a fuzzy balanced scorecard fernando bobillo a,*, miguel delgado a, juan gómez-romero a, enrique lópez b a department of computer science and artificial intelligence, e.t.s.i. informática, university of granada, c. periodista daniel saucedo aranda, granada, spain b economy and business management department, university of león, campus de vegazana s/n, león, spain abstract balanced scorecard is a widely recognized tool to support decision making in business management. unfortunately, current balanced scorecard-based systems present two drawbacks: they do not allow to define explicitly the semantics of the underlying knowledge and they are not able to deal with imprecision and vagueness. to overcome these limitations, in this paper we propose a semantic fuzzy expert system which implements a generic framework for the balanced scorecard. in our approach, knowledge about balanced scorecard vari- ables is represented using an owl ontology, therefore allowing reuse and sharing of the model among different companies. the ontology acts as the basis for the fuzzy expert system, which uses highly interpretable fuzzy if–then rules to infer new knowledge. results are valuable pieces of information to help managers to improve the achievement of the strategic objectives of the company. a main contri- bution of this work it that the system is general and can be customized to adapt to different scenarios. � elsevier ltd. all rights reserved. keywords: knowledge-based systems; expert systems; ontologies semantic web; applied economy; balanced scorecard; fuzzy logic . introduction knowledge management plays a key role in the search for success in the current business world. increasing spe- cialization and complexity of companies has given raise to the necessity of an integral management of own and for- eign resources, which involves and generates huge amounts of valuable data. empresarial intelligence must be conse- quently more a cornerstone of the corporative strategy than simply an amalgam of disperse tools and procedures, if decision processes are expected to be faced with guaran- tees in order to achieve a joint and balanced global performance. balanced scorecard (bsc) (kaplan & norton, ) is a decision support tool at the strategic management level which improves the satisfaction of the strategic objectives. since it was proposed in the early s, it has demon- strated its suitability to assist decision making in management. nevertheless current balanced scorecard-based systems suffer from two problems. firstly, variables which are to be measured have associated vagueness, being much more natural to refer to their values using a linguistic label instead a numerical value as frequently is done. secondly, data do not have an explicit representation of their seman- tics; ad hoc solutions are usually implemented for each problem, making developers duplicate efforts and users cope with their specific details. some solutions have been proposed to the first problem. since fuzzy set theory and fuzzy logic (zadeh, ) have proved to be successful in handling imprecise and vague knowledge, they have been combined with the bsc leads to fuzzy balanced scorecard (see section for details). however, such approaches also leave room for improve- ments in several aspects such as interpretability, modularity and accuracy. on the other hand, to the very well of our knowledge, there has not been any effort in the other direc- tion. thus we have represented balanced scorecard data using an ontology, which allows to add semantics to them - /$ - see front matter � elsevier ltd. all rights reserved. doi: . /j.eswa. . . * corresponding author. tel.: + ; fax: + . e-mail addresses: fbobillo@decsai.ugr.es (f. bobillo), mdelgado@ ugr.es (m. delgado), jgomez@decsai.ugr.es (j. gómez-romero), elopez@ unileon.es (e. lópez). www.elsevier.com/locate/eswa available online at www.sciencedirect.com expert systems with applications ( ) – expert systems with applications mailto:fbobillo@decsai.ugr.es mailto:mdelgado@ mailto:jgomez@decsai.ugr.es mailto:elopez@ making easier knowledge base maintenance as well as reuse of components among different organizations. in this paper we present a new approach to a fuzzy bsc which improves the state of art by extending the number of variables and perspectives. we also present a fuzzy expert system for this fuzzy bsc. its knowledge base relies on an ontology and its inference system derives new knowl- edge from fuzzy rules. the system is general and reusable, so every company can personalize it by providing their own meaning for the linguistic labels defined over the variables (e.g. what they consider a ‘‘high’’ value of some variable) and their own rules. the results of the expert system are highly interpretable pieces of information ready to be incorporated to managers’ decision making processes. the remainder of this paper is structured as follows. section provides some preliminaries on the fundamental theoretical aspects underlying this paper: balanced score- card, fuzzy logic and ontologies. in section we present our fuzzy balanced scorecard, describing the variables which take part in it. implementation details are set out in section . the description of our intelligent system starts by sketching the ontology and then we show how the rule- based engine computes the value of the output variables of the system. section evaluates our proposal with regard to the related work. finally, some conclusions and ideas for future research are drawn in section . . background this section provides some basic background about the topics covered in the paper: section . quickly overviews the original balanced scorecard, section . refreshes the basic ideas in fuzzy sets theory and fuzzy logic, and section . recalls the notion of ontology. . . the balanced scorecard in robert s. kaplan and david p. norton pro- posed the balanced scorecard (bsc) kaplan and norton ( ), a widely recognized tool to support decision mak- ing at the strategic management level which improves the satisfaction of the strategic objectives. the name reflects the objective of maintaining a balance ‘‘between short and long-term objectives, between financial and non-finan- cial measures, between lagging and leading indicators, and between internal and external performance perspectives’’ (kaplan & norton, ). the key innovation of the bsc is, as opposite to tradi- tional approaches which only consider the financial data, to supplement this information with additional non-mone- tary measures. in words of the authors, ‘‘financial measures are inadequate, however, for guiding and evaluating the journey that information age companies must make to create future value through investment in customers, suppliers, employees, processes, technology, and innovation’’. in particular, these authors consider four perspectives: financial perspective obviously, measuring the financial performance of the company, customer perspective, mea- suring the satisfaction of the customers preferences, inter- nal business process perspective, measuring internal business results against measures from financial and cus- tomer perspectives, and innovation and learning perspective, measuring the ability of the company to adapt to changes. a more detailed description of the perspectives is out of the scope of this paper. on the other hand, section depicts the perspectives that we consider. since the apparition of the bsc it has become an impor- tant field of theory and research. many companies have successfully applied this tool and several variations to the original proposal have been investigated (for instance, see section ). . . fuzzy sets and fuzzy logic this section briefly reviews fuzzy sets theory and fuzzy logic; for more details a good reference is (klir & yuan, ). fuzzy set theory and fuzzy logic, proposed by (zadeh, ), are acknowledged as an appropriate formal- ism for capturing imprecise and vague knowledge. while in classical set theory elements either belong to a set or not, in fuzzy set theory elements can belong to a certain degree. more formally, let x be a set of elements. a fuzzy subset a of x, is defined by a membership function laðxÞ which assigns any x x to a value between and . as in the classical case, means no-membership and full member- ship, but now a value between and represents the extent to which x can be considered as an element of x. all crisp set operations are extended to fuzzy sets. the complement, intersection and union set operations are per- formed by a negation function, a t-norm function (typi- cally, the minimum) and t-conorm function (typically, the maximum) respectively. several membership functions can be used in the defini- tion of a fuzzy set. some of the most used are the triangular and the trapezoidal function. a triangular function tria;b;cðxÞ (see fig. a) is defined over the set of non-negative reals rþ [f g with a b c being real numbers. a trap- ezoidal function trza;b;c;dðxÞ is defined over the set of non- negative reals rþ [f g as in fig. b, with a b c d being real numbers. note that a triangular function tria;b;c can be represented using a trapezoidal function trza;b;b;cðxÞ. one of the most important features of fuzzy logic is its ability to perform approximate reasoning (zadeh, ), which involves inference rules with premises, consequences or both of them containing fuzzy propositions. fuzzy rule- based systems have some advantages over other formal- isms: they provide a natural representation for human knowledge as well as a very interpretable model (since the semantics of the rules can easily be understood even for not experts users), are simpler, cheaper and more robust than their crisp versions and, last but not least, have shown to behave very well in practical applications. a fuzzy if–then system consists of a rule base (a set of if–then rules) and a reasoning algorithm performing f. bobillo et al. / expert systems with applications ( ) – https://isiarticles.com/article/ research article expert system for competences evaluation ∘ feedback using fuzzy logic alberto alfonso aguilar lasserre, marina violeta lafarja solabac, roberto hernandez-torres, rubén posada-gomez, ulises juárez-martínez, and gregorio fernández lambert division of research and postgraduate studies, instituto tecnológico de orizaba, avenida instituto tecnológico , colonia emiliano zapata, orizaba,ver, mexico division of research and postgraduate studies, instituto tecnológico superior de misantla, km. . carretera a loma del cojolite, misantla, ver, mexico correspondence should be addressed to alberto alfonso aguilar lasserre; aaguilar@itorizaba.edu.mx received april ; revised june ; accepted june ; published july academic editor: jer-guang hsieh copyright © alberto alfonso aguilar lasserre et al. this is an open access article distributed under the creative commons attribution license, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. performance evaluation (pe) is a process that estimates the employee overall performance during a given period, and it is a common function carried out inside modern companies. pe is important because it is an instrument that encourages employees, organizational areas, and the whole company to have an appropriate behavior and continuous improvement. in addition, pe is useful in decision making about personnel allocation, productivity bonuses, incentives, promotions, disciplinary measures, and dismissals. there are many performance evaluation methods; however, none is universal and common to all companies. this paper proposes an expert performance evaluation system based on a fuzzy logic model, with competences ∘ feedback oriented to human behavior. this model uses linguistic labels and adjustable numerical values to represent ambiguous concepts, such as imprecision and subjectivity. the model was validated in the administrative department of a real mexican manufacturing company, where final results and conclusions show the fuzzy logic method advantages in comparison with traditional ∘ performance evaluation methodologies. . introduction nowadays, labor competences and competence evaluation represent a real challenge for organizations, which emerged in order to assign the right man to the right job. this evaluation method is based on questionnaires that involve fixed scales with specific values, such as %, %, %, %, and %. this kind of evaluation reduces the evaluator opportunity to express points of view and causes a rigid evaluation. fuzzy set theory appears as an important tool to include inaccurate judgments inherent in personnel evaluation process. accord- ing to butkiewicz [ ] fuzzy logic is a very good tool for deci- sion problems, especially when nonprecise or partially precise description is available. although it can be applied with suc- cess in management problems, fuzzy logic is not common in this area. fuzzy logic is an artificial intelligence (ai) technique [ ]. ai comes with the purpose of developing models and programs of the intelligent behavior. one of the approaches of the ai is logic, with the main objective of formalization of natural reasoning. fuzzy logic has two main components: membership functions and fuzzy rules. using them it is possible to move a qualitative to a quantitative description, for example, to rep- resent linguistic expressions as mathematic expressions. this is very useful when it is necessary to model the expertise of a human expert. fuzzy membership functions express the certainty than an element of the universe belongs to a fuzzy set. it represents the degree of truth as an extension of the valuation. degrees of truth are very often confused with probabilities but they are conceptually different because fuzzy truth represents hindawi publishing corporation mathematical problems in engineering volume , article id , pages http://dx.doi.org/ . / / mathematical problems in engineering membership in vaguely defined sets, and not likelihood of an event. these membership functions can take different shapes according to expertise and preferences of the designer. in these membership functions the 𝑥-axis represents the universe of discourse, and the 𝑦-axis represents the degrees of membership in the [ , ] interval. most com- monly used functions are triangular, trapezoidal, gaussian, singleton, gamma, and so forth. membership functions can be expressed as a discrete or continuous function. in other words, 𝜇 𝐴 (𝑋) is a membership function of a set 𝐴, according to the elements of the universe. fuzzy sets are classes of objects with grades of member- ship. each set is characterized by a membership function, which assigns a grade of membership to each object based on a characteristic. when the universe of discourse is continuous and finite, commonly used notation to represent set 𝐴 is 𝐴 = ∫ 𝑥 𝜇 𝐴(𝑥) 𝑥 , ( ) where ≤ 𝜇 𝐴 (𝑥) ≤ . ( ) when the universe of discourse is discrete and finite, fuzzy set 𝐴 is commonly represented as 𝐴 = 𝑛 ∑ 𝑖= 𝜇 𝐴 (𝑥 𝑖 ) 𝑥 𝑖 = 𝜇 𝐴 (𝑥 ) 𝑥 + 𝜇 𝐴 (𝑥 ) 𝑥 + ⋅ ⋅ ⋅ + 𝜇 𝐴 (𝑥 𝑛 ) 𝑥 𝑛 . ( ) for instance, if fuzzy set 𝐴 contains the elements 𝑥 , 𝑥 , 𝑥 , 𝑥 , and 𝑥 with membership degrees of , . , , . , and , respectively, the fuzzy set is expressed as 𝐴 = 𝑥 + . 𝑥 + 𝑥 + . 𝑥 + 𝑥 . ( ) a fuzzy set in a discrete and finite universe of discourse can be represented too as a set of ordered pairs of 𝑥 and its membership degree in 𝐴 as 𝐴 = {(𝑥, 𝜇 𝐴 (𝑥)) | 𝑥 ∈ 𝑋} , ( ) which results in 𝐴 = {(𝑥 , ) , (𝑥 , . ) , (𝑥 , ) , (𝑥 , . ) , (𝑥 , )} . ( ) geometry of fuzzy sets involves three elements: domain, range, and mapping. in this example, geometry of the fuzzy set 𝐴 is domain: {𝑥 , 𝑥 , 𝑥 , 𝑥 , 𝑥 }, range: [ , ], mapping: 𝜇𝐴(𝑥) → [ , ]; { , . , , . , }. graphically, fuzzy set 𝐴 can be expressed as shown in figure . as in classic logic, fuzzy logic uses three basic operations in fuzzy sets: union, intersection, and complement. however, . figure : graphical representation of a fuzzy set. fuzzy sets have certain characteristics that make them differ- ent from classic sets. fuzzy sets have elements with variable membership degrees, which means that an element of the universe of discourse can belong to one or more fuzzy sets, with different membership degrees. the first operation on fuzzy sets is intersection. it is the degree of membership that two fuzzy sets share, that is, is the smallest degree of membership of each element in the fuzzy sets. intersection of two fuzzy sets 𝐴 and 𝐵 is a fuzzy set 𝐴∩𝐵 in the universe of discourse 𝑋, whose function is given by 𝜇 𝐴∩𝐵 (𝑥) = min [𝜇 𝐴 (𝑥) , 𝜇 𝐵 (𝑥)] = ∧ 𝑥 [𝜇 𝐴 (𝑥) , 𝜇 𝐵 (𝑥)], ( ) where 𝐴 ∩ 𝐵 represents the intersection of the fuzzy sets 𝐴 and 𝐵; ∧ represents the minimum operator. the operation union results as the biggest degree of mem- bership of each element in the fuzzy sets, that is, the highest value of the fuzzy values. union of two fuzzy sets 𝐴 and 𝐵 is a fuzzy set 𝐴∪𝐵 in the universe of discourse 𝑋, whose function is given by 𝜇 𝐴∪𝐵 (𝑥) = max [𝜇 𝐴 (𝑥) , 𝜇 𝐵 (𝑥)] = ∨ 𝑥 [𝜇 𝐴 (𝑥) , 𝜇 𝐵 (𝑥)], ( ) where 𝐴 ∪ 𝐵 represents the intersection of the fuzzy sets 𝐴 and 𝐵; ∨ represents the maximum operator. the logic operation complement results as the degree of membership that the fuzzy set needs to reach the unit. the complementary set 𝐴 of a fuzzy set 𝐴 is that whose function is given by 𝜇 − 𝐴 (𝑥) = − 𝜇 𝐴 (𝑥) . ( ) functions that define operations of intersection and union can be generalized using the triangular norm (called 𝑇-norm) and the triangular conorm (called 𝑇-conorm or 𝑆-norm), respectively. mathematical problems in engineering a t-norm operator is a function of two elements 𝑇(⋅, ⋅) that satisfies the following. boundary conditions. this condition implies the generaliza- tion of the classic sets: 𝑇 (𝑎, ) = , 𝑇 (𝑎, ) = 𝑎. ( ) monotonicity. this condition implies that a decrease in the degree of membership for the set 𝐴 or 𝐵 will not produce an increase in the degree of membership of the intersection of the sets 𝐴 and b: 𝑇 (𝑎, 𝑏) ≤ 𝑇 (𝑐, 𝑑) if 𝑎 ≤ 𝑐, 𝑏 ≤ 𝑑. ( ) commutative property. this property indicates that the oper- ator is indifferent to the order of the fuzzy sets that are com- bined: 𝑇 (𝑎, 𝑏) = 𝑇 (𝑏, 𝑎) . ( ) associative property. this property allows calculating the intersection of any number of fuzzy sets, grouped in pairs, regardless of the order of couples: 𝑇 (𝑎, 𝑇 (𝑏, 𝑐)) = 𝑇 (𝑇 (𝑎, 𝑏) , 𝑐) . ( ) in the same way, operator 𝑇-conorm (𝑆-norm) is a func- tion of two elements 𝑆(⋅, ⋅) that satisfies the following. boundary conditions 𝑆 (𝑎, ) = , 𝑆 (𝑎, ) = 𝑎. ( ) monotonicity 𝑆 (𝑎, 𝑏) ≤ 𝑆 (𝑐, 𝑑) if 𝑎 ≤ 𝑐, 𝑏 ≤ 𝑑. ( ) commutative 𝑆 (𝑎, 𝑏) = 𝑆 (𝑏, 𝑎) . ( ) associative 𝑆 (𝑎, 𝑆 (𝑏, 𝑐)) = 𝑆 (𝑆 (𝑎, 𝑏) , 𝑐) . ( ) some interesting contributions of fuzzy logic as a tech- nique to model subjective viewpoints are found in [ ], where the authors firstly considered decision making problems using fuzzy logic. probably, the first attempt to apply fuzzy logic to personnel evaluation was proposed in [ , ]. another approach can be found in [ ]. cannavacciuolo et al. [ ] presented the application of fuzzy set theory to a personnel evaluation procedure. effectiveness of fuzzy concepts and methods depends on the approach used for the analysis of organizational issues. fuzzy set theory allows them to model the weak signals existing in evaluation processes and high- lights part of the tacit knowledge involved in individual judgments. usually, researchers, consultants, and managers use a rather qualitative approach to organizational problems. however the natural language is the preferred instrument to describe the organizational conditions because the shades of meaning and the ambiguity of verbal statements allow the company actors to manage diverging opinions, tensions, and conflicts. these approaches are detailed in [ – ]. on the other hand, the logical-mathematical models tend to represent a world of certainty and coherence where doubts, contradictions, divergences, polysemy, conflicts, and ambigu- ities are usually typified, dissolved, degraded, and linearized. within this same conceptual framework, mathematicians, computer scientists, a.i. researchers, and engineers, in search of formal coherence, quantifiable variables, and efficient algorithms, usually tend to use fuzzy set theory without considering complexity and ambiguity in organizational sit- uations, for example, [ , – ]. there seems to be a growing trend towards the use of systematic procedures in personnel selection. for instance, karsak [ ] introduced a method that integrates decision makers linguistic assessments about subjective factors such as excellence in oral communication skills, personality, lead- ership, and quantitative factors such as aptitude test score within multiple objective programming frameworks. the importance level of each goal is considered by applying the composition operator to the membership function of the goal and the membership function corresponding to its fuzzy priority defined by linguistic variables. kolarik et al. [ ] present an online approach to moni- toring human performance in terms of conditional reliability when a task is performed. unlike traditional human reliability analysis, this approach develops a dynamic model that can cope with constantly changing conditions that affect operator performance. a fuzzy knowledge-based assessment approach is developed in order to deal with uncertainty and subjectivity associated with human performance assessment. podofellini et al. [ ] assess the influence of the failure of the operators to perform one task on the failure probabilities of subsequent tasks with an approach called technique for human error rate prediction (therp) and a fuzzy expert system (fes). other works include the use of fuzzy logic to evolve an optimal and accurate judgment according to the human thinking model and also to mitigate the commonly occurred biases in human recruitment and selection procedures, as seen in [ ]. garćıa et al. [ ] propose, through the use of tools based on fuzzy logic, the evaluation of the impact of training in companies, by applying the reasoning characteristic of fuzzy logic, with the aim of complementing and extending the classical logic. tosti and addison [ ] refer that a poorly designed ∘ feedback system can do more harm than good. the use of commercial ∘ software is not always an option due to spe- cific requirements of each organization. in this way, it may be that people who design ∘ programs are well versed in assessment and measurement technology and woefully mathematical problems in engineering lacking in their understanding of feedback technology. relia- bility of the degree feedback is supported by the number and hierarchy of raters, as referred to in [ ], assuming that personal qualities are developmental goals. therefore, soft- ware has been designed specifically for this study evaluation. this paper is organized as follows. section presents an introduction to the study. section shows some fundamental concepts about competences, performance evaluation ∘ feedback, and competence evaluation methodology; fuzzy logic basis and a proposed fuzzy logic model are shown too. after that, the application of both systems (traditional ∘ feedback system and fuzzy logic ∘ feedback expert system) at the administrative area of a real manufacturing company in the state of veracruz, mexico, is shown. section shows clear and concise results while discussion about the significance of the results is developed. finally, section describes the main conclusions of the study. . expert system for competences evaluation ∘ feedback using fuzzy logic . . competences. personnel appraisal is considered as per- formance evaluation, and it is based on formal evalua- tion programs with reasonable information amount about employees and their job performance. the literature describes several evaluation methods, each with its own advantages and drawbacks, and there is no ideal or universal method for all people, positions, organizations, and situations. the choice will depend on many other aspects such as (i) position, (ii) characteristics to be measured, (iii) organizational culture, (iv) objectives, achieved or to be achieved, (v) circumstantial elements. performance evaluation methods are classified according to the feature they measure characteristics, behaviors, or out- comes, as referred to in [ ]. behavior methods enable the eval- uator to identify how far the employee performance is away from a specific scale. these methods describe what actions should be exhibited during the position performance. it is mainly used to provide development-oriented feedback. according to gomez-mejia et al. [ ], the main advantage in the performance measure adopting a behavior-based approach is that criteria or performance standards are con- crete. behavior scales give employees specific behavior exam- ples that can make them successful (or avoid their success) in their work. if an employee knew the required skills for the position and the corresponding aperture in degrees, it could verify, analyze, and control its own behavior according to the requirements. a competence is an underlying characteristic in the employee related to an effectiveness standard and superior performance in a job or situation, as discussed in [ ]. head evaluation suppliers and customers evaluation peer evaluation subordinate evaluation self-evaluation figure : ∘ evaluation scheme. according to levy-leboyer [ ], individual skills and company competences are closely related. company compe- tences are constituted by the integration and coordination‘of individual skills; however, these competences require an inte- gration and coordination of knowledge and personal qual- ities. individual competences are an individual property. company competences are developed by individuals, but they belong to the company. . . ∘ performance evaluation. ∘ performance evalua- tion is a sophisticated scheme that allows the employee to be evaluated by its surrounding bosses, peers, and subordinates (see figure ). a scheme may include suppliers or customers. ∘performance evaluation can potentially bring a glob- alized diagnosis about the employee performance, allowing the evaluator to compare different opinions about the level of competence expected in the evaluated person and then to take decisions about how to increase the level of compliance of this competences. alles [ ] proposes points for ∘ evaluation. ( ) identify cardinal and individual competences. if the company has implemented a performance evaluation system, the competences will be the same. eventually, it is possible to use a reduced number of competences when using the ∘ evaluation system. ( ) design the tool. questionnaires typically constitute the process support (see figure ). ( ) select evaluators: superiors, partners, internal cus- tomers in other areas, customers, and external suppli- ers. customers can be included or not. it is important to emphasize the fact that assessments are anonymous and that evaluators are chosen by the evaluating person. ( ) launch the evaluation process with stakeholders and evaluators. mathematical problems in engineering figure : questionnaire example. ( ) data processing: most of the time data are processed by external consultants to preserve information con- fidentiality. ( ) communicate ∘ evaluation results to concerned people. ( ) the company will receive a consolidated report. this report will be received only by the employee. in this work, degree feedback methodology proposed by alles [ ] is applied under two different approaches: tra- ditional ∘ feedback and fuzzy logic ∘ feedback. like- wise, there are some other substantial differences. ( ) data processing is performed by two software applications, each designed for its specific process. both applications are able to select evaluators randomly and present questionnaires. from this point, there is another difference: ( ) first application per- forms the evaluation in the traditional degree feedback; the second application is an expert system that uses fuzzy logic into the questionnaires to perform evaluation. the third substantial difference, hence, is that ( ) expert system does not require a human expert, except when they are designed. no external consultant is necessary anymore, since expert system achieves this objective too. . . fuzzy logic basis. fuzzy logic is the mapping from an input measurement space to an output measurement space using linguistic variables. it gives the ability to model impre- cision by incorporating qualitative components into a quan- titative analysis. fuzzy logic systems have a narrow relationship with fuzzy logic concepts such as fuzzy sets and linguistic variables. the most popular fuzzy logic systems are mamdani and takagi- sugeno. fuzzy rules base fuzzifier inference defuzzifier fuzzy input fuzzy output data output mechanism figure : mamdani general fuzzy logic system. mamdani fuzzy systems use components (see figure ). (i) fuzzifier: mamdani system inputs are typically numeric values, coming from some kind of sensor or being results of a process; to be able to operate this value, mamdani systems translate this value into a special value that can be operated by the inference mechanisms. this translation is done by the fuzzifier, which converts numeric values into fuzzy values that represent the level of pertinence of the different variables of the system to the fuzzy sets. (ii) fuzzy inference mechanism: once the fuzzifier has translated the fuzzy values, these have to be processed to generate a fuzzy output. inference mechanism task is to take fuzzy values and generate a fuzzy output based on a fuzzy rules base. (iii) fuzzy rules base is the way in which mamdani fuzzy systems have to represent expertise and linguistic knowledge to solve the issue. it is a set of if-then sentences, containing two parts each: antecedent and conclusion. in a mamdani fuzzy system, antecedent and conclusion are given by linguistic expressions. (iv) defuzzifier: inference system output is a fuzzy output, so it cannot be interpreted by an external element which only could operate numeric data. to make it possible to operate this data, output is translated to numeric format, and this task is done by the defuzzi- fier, using one of different procedures such as gravity center or averaged centers. fuzzy logic uses certain essential components to achieve its purpose. imprecision. often the same term is used to describe impreci- sion and uncertainty in only slightly related areas of measure- ment. imprecision in measurement is associated with a lack of knowledge. imprecision as a probability form is associated with uncertainty about the future event occurrence. impreci- sion in description, the imprecision type addressed by fuzzy logic, is connected with intrinsic or built-in imprecision that belongs to the event itself. fuzzy logic addresses the issues associated with an intrin- sic imprecision rather than those directly concerned with measuring devices failures in the measurements accuracy. mathematical problems in engineering intrinsic imprecision is associated with a phenomenon prop- erties description and not with properties measurement using some external device. ambiguity. there are close semantic relationships between the ambiguity idea and fuzziness; in fact, some fuzzy states can be highly ambiguous. ambiguity connotes the property to have several but plausible and reasonable interpretations. these interpretations can have different belief states. ambiguity in meaning is a common occurrence in natural languages. likelihood and ambiguity. fundamentally, the basic confu- sion between fuzzy logic and probability arises from the idea that they measure the same kind of uncertainty. in strictly semantic, as well as mechanistic, the two forms of uncertainty are different. propositions in probability address the likeli- hood of an outcome for some discrete event. the event out- come either happens or does not happen. propositions in fuzzy logic concern the degree to which an event occurred. while a probability outcome happens unequivocally, a fuzzy event occurrence may involve some degree of ambiguity or uncertainty. fuzzy sets components. gregory [ ] indicates that the fuzzy logic has two main components: membership functions and fuzzy rules. when using these components it is possible to move the experiences and human preferences from a qualitative description to a quantitative description. membership fuzzy functions can take different figures and forms, according the designer experiences and pref- erences. typical functions are triangular, trapezoidal, s, gamma, gaussian, and exponential. on the other hand, the fuzzy rules are written as if-then couples and reported in tabular form. the four basic ways in which the fuzzy rules can be achieved are expert experiences and engineering knowledge, human behaviors, models based on a fuzzy system, and learn- ing processes. these methods are not necessarily mutually exclusive. membership functions. in classical set theory, something is completely included or not. this situation can be described by assigning a value of one to all the elements included in the set and the value of zero to the ones not included in it. the func- tion that assigns these values is called “membership function.” the fuzzy sets allow to describe the degree of membership of the object to the concept given by the labels, and allow to assign values between zero and one to the membership function (see figure ). mathematical features of fuzzy sets. main characteristics of the fuzzy sets are height, support, cutoff level-𝛼, and nucleus. height. it is the highest degree of membership of the elements of the set; that is, height (𝐴) = max {ℎ | ℎ = 𝜇 𝐴 (𝑥) , 𝑥 ∈ 𝑋} . ( ) when the height of a fuzzy set is equal to it is said that it is a normalized fuzzy set. d eg re e of m em be rs hi p 𝜇 (x ) connects domain rk with 𝜇(x)i r rn figure : general structures in a fuzzy set. a b c d . 𝜇 u figure : trapezoidal type membership function. support. it is the number of elements whose degree of membership is not zero; that is, sup (𝐴) = {𝑥 | 𝜇 𝐴 (𝑥) > 𝑂, 𝑥 ∈ 𝑋} . ( ) cutoff level-𝛼. it is the set of elements of 𝑋 with a minimum degree 𝛼; that is, 𝐴 𝛼 = {𝑥 | 𝜇 𝐴 (𝑥) ≥ 𝛼, 𝑥 ∈ 𝑋} . ( ) nucleus. it is the set of elements of 𝑋 with a degree of membership equal to ; that is, nucleus (𝐴) = {𝑥 | 𝜇 𝐴 (𝑥) = , 𝑥 ∈ 𝑋} . ( ) inclusion functions in fuzzy sets. there are standard families for inclusion functions; the most frequent ones are trape- zoidal, singleton, triangular, 𝑆, exponential, and 𝜋 type. trapezoidal functions are defined by four points: 𝑎, 𝑏, 𝑐, and 𝑑 (see figure ). this function is zero for values lower than “𝑎” and higher than “𝑑” and one between “𝑏” and “𝑐” and takes values in range [ , ] between “𝑎” and “𝑏” and between mathematical problems in engineering “𝑐” and “𝑑.” it is used in simple fuzzy systems, since it allows defining a fuzzy set with little information and computing membership function values in a simple way. this function is common for microprocessor based sys- tems since it can be encoded in a similar format as𝑆 functions, 𝜋 functions, and triangular and singleton functions (e.g., if points 𝑏 and 𝑐 are combined the result is a triangular function). trapezoidal function is defined as follows (see ( )): 𝑆 (𝑢; 𝑎, 𝑏, 𝑐, 𝑑) = { { { { { { { { { { { { { { { { { { { { { { { { { , 𝑢 < 𝑎, ( 𝑢 − 𝑎 𝑏 − 𝑎 ) , 𝑎 ≤ 𝑢 ≤ 𝑏, , 𝑏 ≤ 𝑢 ≤ 𝑐, ( 𝑑 − 𝑢 𝑑 − 𝑐 ) , 𝑐 ≤ 𝑢 ≤ 𝑑, , 𝑢 > 𝑑. ( ) trapezoidal functions are suitable to model properties in a range of values, stages, or levels (e.g., young, adult, elder, etc.). modeling a triangular function can be done through the 𝑏 = 𝑐 simplification. for an 𝑆 function and singleton types (but not soft), 𝑐 = 𝑑 = max(𝑈) and 𝑎 = 𝑏 = 𝑐 = 𝑑 transformations can be applied, respectively. triangular function (𝑇) can be defined as indicated in 𝑇 (𝑢; 𝑎, 𝑏, 𝑐) = { { { { { { { { { { { { { { { { { , 𝑢 < 𝑎, 𝑢 − 𝑎 𝑏 − 𝑎 , 𝑎 ≤ 𝑢 ≤ 𝑏, 𝑐 − 𝑎 𝑐 − 𝑏 , 𝑏 ≤ 𝑢 ≤ 𝑐, , 𝑢 > 𝑐. ( ) 𝑇 functions (see figure ) are appropriate for modeling properties with an inclusion value different from zero and for a narrow range of values around point 𝑏. linguistic variables. linguistic variables take values from natural language, for example, much, little, positive, negative, and so forth. these words are considered as labels within the fuzzy set theory. even though linguistic variables aim to assign labels as variable values taken from natural language words, they will be able to assign numerical values too. then, in the expression “temperature is cold,” the variable “temperature” must be seen as a linguistic variable, since the value cold is assigned as a fuzzy set. however, this variable can also take numerical values such as “temperature is ∘c.” fuzzy rules. fuzzy rules combine one or more input fuzzy sets, called premises, and associate them with an outcome fuzzy set, called consequence. the fuzzy set premise is asso- ciated using and, or, and so forth operators. defuzzification process. there are two common methods of defuzzification process: gravity center (centroid) and maxi- mum output. as is shown in figure , both techniques pro- duce different results [ ]. a b c . 𝜇 u . figure : type t (triangular) function. both techniques produce reasonable results when they are applied in specific fuzzy models. gravity center is the most common method, because it combines evidence about rules and response fields are pondered by the total true degree. gravity center is, essentially, the weighted average of the output membership function. centroid computation. centroid technique finds the balance point solution in fuzzy zone using the weighted average in the fuzzy region. arithmetically, the procedure is formulated by r = ∑ 𝑛 𝑖= 𝑑 𝑖 ⋅ 𝜇 𝐴 (𝑑 𝑖 ) ∑ 𝑛 𝑖= 𝜇 𝐴 (𝑑 𝑖 ) , ( ) where 𝑑 is the 𝑖th domain value and 𝜇(𝑑) is the true membership value at this point. centroid or defuzzification with moments finds a point that represents the fuzzy set gravity center. . . fuzzy logic personnel evaluation model. a competence performance evaluation involves subjective viewpoints and evaluator preferences are reflected at the evaluation moment. very often, evaluators express their perceptions in natural language terms. nevertheless, as mentioned above, evalu- ation questionnaires constitute the performance evaluation support, and they are based on punctual values that do not reflect or approximate real viewpoints. therefore, it is desirable to have a flexible evaluation tool that facilitates imprecision, ambiguity, and subjectivity han- dling. in this purpose, fuzzy sets allow a suitable treatment. the fuzzy logic model proposed is constituted as follows. linguistic variables. three linguistic variables have been defined, “scale,” “frequency,” and “required level.” “scale” variable refers to the percentage (assigned by the evaluator) indicating how well the employee behavior matches the competence definition. “frequency” refers to the percentage obtained when the evaluator answers a question and ”rethinks” his/her evalua- tion determining the number of times the behavior is mani- fested. mathematical problems in engineering center of gravity numbers d eg re e of m em be rs hi p 𝜇 (x ) (a) maximum of output numbers d eg re e of m em be rs hi p 𝜇 (x ) (b) figure : defuzzification common methods. needs significant development needs development competent highly competent role model a b c d e figure : fuzzy sets for linguistic variable “scale.” “required level” refers the percentage expected by the organization; the individual must cover conduct according the competence definition. output variables. they refer to the fit conduct qualification corresponding to the competence definition given in percent- age. under this consideration there are three cases, “needs to improve,” “satisfies,” and “exceeds.” fuzzy sets representation. this model uses triangular inclu- sion functions to represent linguistic variables; the output variable was modeled with trapezoidal functions. according to the expert, five fuzzy sets define the variable “scale” possible values (see figure ), which are (i) “needs significant development,” (ii) “needs development,” (iii) “competent,” (iv) “highly competent,” (v) “role model.” likewise, there are four fuzzy sets to define the variable “frequency” possible values (see figure ), which are (i) “occasionally,” (ii) “half time,” (iii) “frequent,” (iv) “always.” the “required level” variable was modeled through three possible fuzzy sets (see figure ), called (i) “low,” (ii) “average,” (iii) “high.” the output variable has three cases which allow defining fuzzy sets (see figure ); these cases are (i) “requires improvement,” (ii) “complies,” (iii) “exceeds.” mathematical problems in engineering occasionally halftime frequent always a b c d figure : fuzzy sets for linguistic variable “frequency.” low average high a b c figure : fuzzy sets for linguistic variable “required level.” requires improvement complies exceeds . . a b c d figure : fit qualification. fuzzy rules. fuzzy rules must consider all the combinations among input variables sets. each combination will be associ- ated with an output variable fuzzy set. sixty fuzzy rules were created in this case, according to the opinion of an expert, who rigorously analyzed each set of inputs, determining the level of performance produced by each rule (see table ). fuzzification. the fuzzification process includes membership functions calculation for input variables and, then, uses the minimum-maximum criterion for variables activation. defuzzification. the defuzzification process will allow getting the final fit qualification; the proposed method is the gravity center (centroid) including the following four steps: ( ) divide total area in partial areas, ( ) calculate partial areas value, ( ) calculate each partial area centroid, ( ) calculate total centroid. . . computational experiments. this section describes the use of the methodology using both traditional ∘ feedback and ∘ fuzzy logic feedback. . . . application of the expert system in a real company. the use of both methods was carried out in the administrative area of a mexican manufacturing company, located in the state of veracruz, mexico, where administrative procedures require staff to be evaluated based on their performance annually. a branch of the organization chart, consisting of five different but related positions was selected to perform the ∘ feedback evaluation process (see figure ). . . . traditional ∘ methodology ( ) identify cardinal and individual competences. to evaluate a position we will consider five cardinal com- petences and eight specific competences (see table ). ( ) design the tool. this evaluation includes four ques- tionnaires and each questionnaire includes five ques- tions. ( ) select an evaluator. evaluation includes self-evalua- tion, boss evaluation, and peers evaluation. to choose evaluators and launch the process we will design and use the software; in this sense, the selection procedure is random. ( ) execute the evaluation process with concerned people and evaluators. as an example, table shows self- evaluation values for cardinal competences. mathematical problems in engineering ta bl e : fu zz y ru le ss um m ar y. sc a le n ee ds si gn ifi ca nt de ve lo pm en t n ee ds de ve lo pm en t c om pe te nt h ig hl y co m pe te nt r ol e m od el fr eq ue nc y fr eq ue nc y fr eq ue nc y fr eq ue nc y fr eq ue nc y occasional halftime frequent always occasional halftime frequent always occasional halftime frequent always occasional halftime frequent always occasional halftime frequent always r eq ui re d le ve l lo w c ( ) c ( ) i( ) i( ) c ( ) c ( ) i( ) i( ) c ( ) c ( ) e ( ) e ( ) c ( ) c ( ) e ( ) e ( ) e ( ) c ( ) e ( ) e ( ) a ve ra ge c ( ) i( ) i( ) i( ) c ( ) i( ) i( ) i( ) i( ) c ( ) c ( ) c ( ) c ( ) c ( ) c ( ) e ( ) c ( ) c ( ) c ( ) e ( ) h ig h i( ) i( ) i( ) i( ) i( ) i( ) i( ) i( ) i( ) i( ) c ( ) c ( ) i( ) i( ) c ( ) c ( ) i( ) i( ) c ( ) c ( ) i: re qu ir es im pr ov em en t. c :c om pl ie s. e: ex ce ed s. mathematical problems in engineering operational human resources coordinator superintendent of operational assessment specialist of assessment and control occupational health specialist principal analyst of occupational health superintendent of organization and integration labor productivity specialist labor productivity specialist studies and integration specialist superintendent of professional competitiveness professional competitiveness specialist assessment and recruitment specialist superintendent of operation and services operation and services specialist operations and services analyst superintendent of labor regulations labor regulation analyst selected positions figure : organizational chart. table : cardinal and specific competences. cardinal competences specific competences ( ) work quality ( ) planning and organization skills ( ) tolerance to work under pressure ( ) customer orientation ( ) communication skills ( ) productivity ( ) contact modalities ( ) technical credibility ( ) analytical skills ( ) innovation ( ) empowerment ( ) collaboration ( ) commitment level-personal discipline-productivity ( ) data processing: software specifically designed for data processing to obtain the competence level scale value and frequency value was used; then we will aver- age the results for each question (see table ). finally, averages are compared with the required level. the same procedure will be used for specific competences treatment. ( ) reports will be delivered only to the evaluated person. comments and graphs will be printed in two reports: the first one will include cardinal competences and will be given to the company, while the second report will include specific competences and will be given to the employee. ( ) communicate the ∘ evaluation results to the con- cerned people. . . . fuzzy logic model implementation. ( ) choose an organ- izational department. ( ) choose a post in the selected department. ( ) choose a required competence in the post. ( ) select one question in the questionnaire that evaluates the competence. ( ) evaluator assigns values for input variables “scale” and “frequency,” for example, scale = % frequency = % the required level assigned by the company = %. ( ) fuzzification process: compute the variable member- ship values corresponding to the variable scale: 𝜇competent (scale) = { { { { { { { { { { { { { { { { { { { ; 𝑊 ≤ 𝐵 󸀠 , − 𝐶 󸀠 − 𝑊 𝐶 󸀠 − 𝐵 󸀠 ; 𝐵 󸀠 < 𝑊 ≤ 𝐶 󸀠 , − 𝑊 − 𝐶 󸀠 𝐷 󸀠 − 𝐶 󸀠 ; 𝐶 󸀠 < 𝑊 < 𝐷 󸀠 . , ; 𝐷 󸀠 ≤ 𝑊 , 𝜇highly⋅competent (scale) = { { { { { { { { { { { { { { { { { { { ; 𝑊 ≤ 𝐶 󸀠 , − 𝐷 󸀠 − 𝑊 𝐷 󸀠 − 𝐶 󸀠 ; 𝐶 󸀠 < 𝑊 ≤ 𝐷 󸀠 . , − 𝑊 − 𝐷 󸀠 𝐸 󸀠 − 𝐷 󸀠 ; 𝐷 󸀠 < 𝑊 < 𝐸 󸀠 , ; 𝐸 󸀠 ≤ 𝑊 . ( ) mathematical problems in engineering ta bl e : se lf- ev al ua tio n va lu es ,c ar di na lc om pe te nc es . (a ) ∘ ev al ua tio n se lf- ev al ua tio n c ar di na lc om pe te nc ie s w or k qu al ity r eq ui re d le ve l: to le ra nc e to w or k un de rp re ss ur e r eq ui re d le ve l: c om m un ic at io n r eq ui re d le ve l: c on ta ct m od al iti es r eq ui re d le ve l: a na ly tic sk ill s r eq ui re d le ve l: sc al e fr eq ue nc y sc al e fr eq ue nc y sc al e fr eq ue nc y sc al e fr eq ue nc y sc al e fr eq ue nc y q ue s. q ue s. q ue s. q ue s. q ue s. (b ) w he re sc al e fr eq ue nc y % :a % :a lw ay s % :b % :f re qu en t % :c % :h al ft im e % :d % :o cc as io na l % :n o % :n ev er mathematical problems in engineering ta bl e : se lf- ev al ua tio n, ca rd in al co m pe te nc es tr ea tm en t. (a ) ∘ ev al ua tio n se lf- ev al ua tio n c ar di na lc om pe te nc ie s w or k qu al ity r eq ui re d le ve l: pr es su re to le ra nc e r eq ui re d le ve l: c om m un ic at io n r eq ui re d le ve l: c on ta ct m od al iti es r eq ui re d le ve l: a na ly tic sk ill s r eq ui re d le ve l: sc al e fr eq ue nc y pr od uc t sc al e fr eq ue nc y pr od uc t sc al e fr eq ue nc y pr od uc t sc al e fr eq ue nc y pr od uc t sc al e fr eq ue nc y pr od uc t q ue s. . q ue s. . . q ue s. q ue s. . q ue s. . . . a ve ra ge . . . . . (b ) w he re sc al e fr eq ue nc y % :a % :a lw ay s % :b % :f re qu en t % :c % :h al ft im e % :d % :o cc as io na l % :n o % :n ev er mathematical problems in engineering compute the variable membership values corresponding to the variable frequency: 𝜇frequent (frequency) = { { { { { { { { { { { { { { { { { { { ; 𝑊 ≤ 𝐵 , − 𝐶 − 𝑊 𝐶 − 𝐵 ; 𝐵 < 𝑊 ≤ 𝐶 , − 𝑊 − 𝐶 𝐷 − 𝐶 ; 𝐶 < 𝑊 < 𝐷 . , ; 𝐷 ≤ 𝑊 . 𝜇always (frequency) = { { { { { { { { { ; 𝑊 ≤ 𝐶 , − 𝐷 − 𝑊 𝐷 − 𝐶 ; 𝐶 < 𝑊 ≤ 𝐷 . , ; 𝐷 ≤ 𝑊 . ( ) compute the variable membership values corresponding to the variable required level: 𝜇average (requiredlevel) = { { { { { { { { { { { { { { { { { { { ; 𝑊 ≤ 𝐴 󸀠󸀠 , − 𝐵 󸀠󸀠 − 𝑊 𝐵 󸀠󸀠 − 𝐴 󸀠󸀠 ; 𝐴 󸀠󸀠 < 𝑊 ≤ 𝐵 󸀠󸀠 , − 𝑊 − 𝐵 󸀠󸀠 𝐶 󸀠󸀠 − 𝐵 󸀠󸀠 ; 𝐵 󸀠󸀠 < 𝑊 < 𝐶 󸀠󸀠 , ; 𝐶 󸀠󸀠 ≤ 𝑊 . ( ) the fuzzification process continues with variable activa- tion, performed according to the min.-max. criterion (see ( ), ( ), ( ), and ( )); this procedure allows identifying the membership values that will appear in the defuzzification process. rule is 𝜇 𝐶∩𝐹∩𝑀 ( , , ) = min {𝜇 𝐶 ( ) , 𝜇 𝐹 ( ) , 𝜇 𝑀 ( )} = min { . , . , . } = . . ( ) rule is 𝜇 𝐶∩𝑆∩𝑀 ( , , ) = min {𝜇 𝐶 ( ) , 𝜇 𝑆 ( ) , 𝜇 𝑀 ( )} = min { . , . , . } = . . ( ) rule is 𝜇 𝐴∩𝐹∩𝑀 ( , , ) = min {𝜇 𝐴 ( ) , 𝜇 𝐹 ( ) , 𝜇 𝑀 ( )} = min { . , . , . } = . . ( ) table : membership degree for activated variables. activated variable set, output variable membership degree requires improvement . complies . complies . complies . table : membership degree for activated variables. activated variable fuzzy set membership degree — exceeds , , complies . requires improvement . rule is 𝜇 𝐴∩𝑆∩𝑀 ( , , ) = min {𝜇 𝐴 ( ) , 𝜇 𝑆 ( ) , 𝜇 𝑀 ( )} = min { . , . , . } = . . ( ) the rules , , , and are the activated variables; they have correspondence with a set in the output variable (see table ). three activated variables are in the “obey” set; it is nec- essary to use the min.-max. criterion to select one value (see ( )): 𝜇 𝐶 = ( , , ) = ∨ {∧ {𝜇 𝐶 ( ) , 𝜇 𝑆 ( ) , 𝜇 𝑀 ( )} , ∧ {𝜇 𝐴𝐶 ( ) , 𝜇 𝐹 ( ) , 𝜇 𝑀 ( )} , ∧ {𝜇 𝐴𝐶 ( ) , 𝜇 𝑆 ( ) , 𝜇 𝑀 ( )}} = . . ( ) the last membership degrees and activated variables are summarized as follows (see table ). ( ) defuzzification process: this procedure includes five steps. divide the total area in partial areas (see figure ). compute the partial area values (see ( )): (𝐴i) = [ . ] ( . − ) = . , (𝐴ii) = [ . ] [ − . ] ( . − ) = . , (𝐴iii) = [ . ] ( . − ) = , (𝐴iv) = [ . ] ( . − . ) = , (𝐴v) = [ . ] [ − . ] ( − . ) = . , (𝐴vi) = [ . ] ( − . ) = . . ( ) mathematical problems in engineering requires improvement complies exceeds . . a b c d . . i ii iii iv v vi figure : partial areas. compute each partial area centroid (see ( )): centroid (𝐴i) = + ( . − ) ( − . ) = , centroid (𝐴ii) = . − ( . − ) [ . ] = . , centroid (𝐴iii) = . − ( . − ) [ . ] = . , centroid (𝐴iv) = . + . = , centroid (𝐴v) = . + ( − . ) ( − . ) = . , centroid (𝐴vi) = − ( − . ) [ . ] = . ( ) calculate the total centroid (see ( )): cet = (( . ) ( ) + ( . ) ( . ) + ( ) ( . ) + ( ) ( ) + ( . ) ( . ) + ( . ) ( )) × ( . + . + + + . + . ) − = . . = . . ( ) ( ) competences questionnaires have more than one question. in this sense, each questionnaire will have as many fuzzy treatments as questions. at the end of the process, an average qualification for each competence will be obtained. ( ) evaluation software and evaluation report. intelli- gent evaluation systems have questionnaires that emulate human thought since it gives the option to answer questions using linguistic labels (see figure ). once the evaluator has selected a label, it selects a numerical value. final stage includes an evaluation report,which includes a summarized table with average qualifications for each competence. figure : evaluation questionnaire. table : cardinal competences and required level. cardinal competences required level work quality % pressure tolerance % communication % contact modalities % analytic skill % . results . . ∘ methodology application. a complete application includes four-position analysis (see figure ); five cardinal competences were developed using the job description (see table ). the first position has nine specific competences, while the second position has ten specific competences (see table ), the third position has eight competences, and finally the fourth position has ten competences. required levels were established with the company. applying the complete traditional methodology ∘ to the first position, we concluded that the company must improve “work quality” and “contact modalities,” since these competences obtained the major difference with the required level (see table ). the individual report includes specific competences sum- mary. mathematical problems in engineering table : specific competences and required level. specific competences position level i required level position level ii required level ( ) work team % ( ) planning and organization skills % ( ) negotiation % ( ) negotiation % ( ) personnel development % ( ) initiative autonomy % ( ) leadership % ( ) leadership % ( ) frankness-trustworthy-integrity % ( ) frankness-trustworthy-integrity % ( ) commitment % ( ) empowerment % ( ) collaboration % ( ) strategic thinking % ( ) coaching % ( ) commitment level-personal discipline-productivity % ( ) decision making % ( ) making decisions % ( ) human resources strategic development % table : cardinal competences report. cardinal competences work quality pressure tolerance communication contact modalities analytic skill self-evaluation . . . . . peer . . . . . head . . . . . ∘ weighted . . . . . required level . . . . . difference: required level, ∘ weighted . ∗ . . . ∗ . ∗competences with major difference. table : specific competences report. specific competences planning and organization ability client orientation productivity credibility technique innovation empowerment collaboration commitment level-personal discipline- productivity self-evaluation . . . . . . . . peer . . . . . . . . boss . . . . . . . . ∘ weighted . . . . . . . . required level . . . . . . . . difference: required level, ∘ weighted . . . . . . . ∗ . ∗ ∗competences with major difference. applying the complete traditional methodology ∘ to the same post, we concluded that an employee must improve “collaboration” and “commitment level-personal discipline- productivity,” since these competences obtained the biggest difference with the required level (see table ). . . fuzzy logic model application. applying the full fuzzy logic model evaluation to the first position, we concluded that the company must improve “communication” and “contact modalities,” since these competences obtained the lower final qualification. employee must focus on “credibility technique” and “commitment level-personal discipline-productivity” (see figure ). . discussion traditional methodology ∘ indicates competences that the company and the employee must care. fuzzy logic ∘ methodology indicates the competences on which the com- pany and the employees must focus their efforts but in addi- tion gives a qualification. this qualification can be arranged and the lowest value will indicate which competences must be immediately attended. according to table , the traditional ∘ and fuzzy logic methodologies conclude that the company must attend the cardinal competence “contact modalities” and the employee must focus on the specific competence “commitment level- personal discipline-productivity.” however, fuzzy logic mathematical problems in engineering table : comparison results. competences classification traditional ∘ methodology fuzzy logic ∘ expert system cardinal ∗work quality ∗contact modalities ∗communication: . ∗contact modalities: . specific ∗collaboration ∗commitment level-personal discipline- productivity ∗technical credibility: . ∗commitment level-personal discipline-productivity: . figure : qualifications summary. method indicates that “contact modalities” must be attended after “communication,” and “commitment level-personal dis- cipline-productivity” must be attended after “technical credibility.” the traditional ∘ system involves only two factors, “scale” and “frequency,” and these are compared with the third factor “required level.” on the other hand, fuzzy logic model can involve three factors like input variables. evaluation questionnaires of the traditional ∘method- ology are filled with five possible percentages ( , , , , and ). on the other hand, questionnaires in fuzzy logic expert system have adjustable values scales, allowing improved flexibility for the evaluator. thus, the main advantage in fuzzy logic expert system is the human thinking simulation, assigning labels as a qualifi- cation, and allowing subjective and ambiguity treatment. . conclusions competences performance evaluation ∘ is a complete sys- tem since it involves different viewpoints to appraise per- sonnel performance. it allows better interpretations, since the evaluation responsibility falls in different evaluators. this kind of evaluation facilitates the competence concept comprehension, and required levels are assimilated. never- theless, it is a rigid system because the questionnaire filling procedure is strict, since these were designed to assign fixed values. fuzzy logic competences evaluation expert system includes complex analysis, due to identification and modeling input and output variables. however, its main advantage is the ambiguity and subjectivity handling, since the evaluator can assign words to stand a qualification. this system is flexible because numerical adjustable values can be assigned to behaviors. graphical interpretation helps to obtain suitable feedback. even more, final processing reports can be obtained easier and faster. therefore, it represents an excellent tool for competences monitoring, given the importance that the staff appraisal process has for human resource management, in the areas of recruitment and selection, job evaluation, identification of training needs, and so forth, and the value that results from having nonsubjective and bias-free assessments. the application of an expert system to performance eval- uation in a mexican manufacturing company allows knowing its effectiveness against traditional techniques. this work brings innovative contributions to soft com- puting and human resources management solutions, finding new ways to apply artificial intelligence techniques by means of computer applications to processes that typically were performed by humans. conflict of interests the authors declare that there is no conflict of interests regarding the publication of this paper. references [ ] b. s. butkiewicz, “selection of staff for enterprise using fuzzy logic,” in ieee international conference on systems, man and cybernetics, vol. , warsaw university of technology, warsaw, poland, . [ ] e. gómez, modelo difuso de planeación agregada [doctoral thesis], centro de ingenieŕıa y desarrollo industrial, querétaro, méxico, . [ ] l. a. zadeh, “making computers think like people,” ieee spectrum, vol. , no. , pp. – , . [ ] a. cannavacciuolo, g. capaldo, a. ventre, and g. zollo, “link- ing the fuzzy set theory to organizational routines. a study in personnel evaluation in a large company,” in proceedings of the nd ieee international conference on fuzzy systems, pp. – , san francisco, calif, usa, april . [ ] a. cannavacciuolo, g. capaldo, a. ventre, and g. zollo, “an approach to the evaluation of human resourses by using fuzzy set theory,” in proceedings of the rd ieee conference on fuzzy systems, orlando, fla, usa, . [ ] j. stahl, “a method for personnel selection in concurrent engi- neering using fuzzy sets,” in fuzzy sets in engineering design and configuration, h.-j. sebastian and e. k. antonsson, eds., kluwer academic publishers, boston, mass, usa, . [ ] c. argyris and d. a. schon, organizational learning. a theory of action perspective, addison-wesley, reading, mass, usa, . mathematical problems in engineering [ ] k. s. cameron, “effectiveness as paradox: consensus and con- flict in conceptions of organizational effectiveness,” journal of management science, vol. , no. , pp. – , . [ ] d. j. orton and k. e. weik, “loosely coupled system: a recon- ceptualization,” academy of management review, vol. , no. , pp. – , . [ ] c. r. schwenk, “cognitive simplification processes in strategic decision-making,” strategic management journal, vol. , pp. – , . [ ] k. e. weick, “educational organizations as loosely coupled sys- tems,” administrative science quarterly, vol. , pp. – , . [ ] c. carisson, “on the relevance of fuzzy sets in management science methodology,” in fuzzy sets: theory and applications to policy analysis and information system, p. p. wang and s. k. chang, eds., plenum press, new york, ny, usa, . [ ] b. r. gaines, l. a. zadeh, and h. zimmermann, “fuzzy sets and decision analysis—a perspective,” in fuzzy sets and decision analysis, h. j. zimmermann and etal, eds., north-holland, amsterdam, netherlands, . [ ] r. m. tong and p. p. bonissone, “linguistic solutions to fuzzy decision problems,” in studies in the management science, h. j. zimmermann, ed., pp. – , . [ ] r. r. yager, “fuzzy thinking as quick and efficient,” cybernetica, vol. , no. , pp. – , . [ ] p. l. yu, “dissolution of fuzziness for better decision-perspec- tive and techniques,” in fuzzy sets and decision analysis, j. h. zimmermann et al., ed., north-holland, amsterdam, the netherlands, . [ ] s. french, “fuzzy decision analysis: some criticism,” in fuzzy sets and decision analysis, h. j. zimmermann, b. r. gaines, and l. zadeh, eds., amsterdam, the netherlands, north holland publisher, . [ ] l. a. zadeh, “outline of new approach to the analysis of complex systems and decision processes,” ieee transactions on systems, man, and cybernetics, vol. , pp. – , . [ ] e. e. karsak, “a fuzzy multiple objective programming approach for personnel selection,” in proceedings of the ieee international conference on systems, man, and cybernetics, vol. , pp. – , nashville, tenn, usa, october . [ ] w. j. kolarik, j. c. woldstad, s. lu, and h. lu, “human performance reliability: on-line assessment using fuzzy logic,” iie transactions, vol. , no. , pp. – , . [ ] l. podofellini, v. dang, e. zio et al., “using expert models in human reliability analysis—a dependence assessment method based on fuzzy logic,” risk analysis, vol. , no. , pp. – , . [ ] k. c. mittal, a. k. goel, and p. mohindru, “fuzzy multi-criteria decision making (mcdm) in human resource selection proce- dure—a case study of indian it industry,” bvimr management edge, vol. , no. , pp. – , . [ ] r. e. garćıa, g. felix benjamin, and r. bello perez, “impact assessment of training with fuzzy logic,” ingeniare. revista chilena de ingenieŕıa, vol. , no. , pp. – , . [ ] d. t. tosti and r. m. addison, “ -degree feedback: going around in circles?” performance improvement, vol. , no. , pp. – , . [ ] r. hensel, f. meijers, r. van der leeden, and j. kessels, “ degree feedback: how many raters are needed for reliable ratings on the capacity to develop competences, with personal quali- ties as developmental goals?” international journal of human resource management, vol. , no. , pp. – , . [ ] l. r. gomez-mejia, d. b. balkin, and r. l. cardy, managing human resources, prentice hall, saddle river, nj, usa, nd edition, . [ ] m. spencer lyle and m. spencer signe, competence at work: models for superior performance, . [ ] c. levy-leboyer, “evaluación del personal: los métodos a ele- gir,” ediciones dı́az de santos, . [ ] m. a. alles, gestión por competencias el diccionario, granica, buenos aires, argentina, nd edition, . [ ] c. gregory, dynamic economics: optimization by the lagrange method, oxford university press, . [ ] e. cox, the fuzzy systems handbook: a practitioner’s guide to building, using, and maintaining fuzzy systems, ap. profes- sional, . submit your manuscripts at http://www.hindawi.com hindawi publishing corporation http://www.hindawi.com volume mathematics journal of hindawi publishing corporation http://www.hindawi.com volume mathematical problems in engineering hindawi publishing corporation http://www.hindawi.com differential equations international journal of volume applied mathematics journal of hindawi publishing corporation http://www.hindawi.com volume probability and statistics hindawi publishing corporation http://www.hindawi.com volume journal of hindawi publishing corporation http://www.hindawi.com volume mathematical physics advances in complex analysis journal of hindawi publishing corporation http://www.hindawi.com volume optimization journal of hindawi publishing corporation http://www.hindawi.com volume combinatorics hindawi publishing corporation http://www.hindawi.com volume international journal of hindawi publishing corporation http://www.hindawi.com volume operations research advances in journal of hindawi publishing corporation http://www.hindawi.com volume function spaces abstract and applied analysis hindawi publishing corporation http://www.hindawi.com volume international journal of mathematics and mathematical sciences hindawi publishing corporation http://www.hindawi.com volume the scientific world journal hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume algebra discrete dynamics in nature and society hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume decision sciences advances in discrete mathematics journal of hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume stochastic analysis international journal of a belief rule based expert system to assess tuberculosis under uncertainty j med syst ( ) : doi . /s - - - mobile & wireless health a belief rule based expert system to assess tuberculosis under uncertainty mohammad shahadat hossain · faisal ahmed · fatema-tuj-johora · karl andersson received: november / accepted: january / published online: january © the author(s) . this article is published with open access at springerlink.com abstract the primary diagnosis of tuberculosis (tb) is usually carried out by looking at the various signs and symptoms of a patient. however, these signs and symptoms cannot be measured with % certainty since they are associated with various types of uncertainties such as vague- ness, imprecision, randomness, ignorance and incomplete- ness. consequently, traditional primary diagnosis, based on these signs and symptoms, which is carried out by the physicians, cannot deliver reliable results. therefore, this article presents the design, development and applications of a belief rule based expert system (brbes) with the ability to handle various types of uncertainties to diag- nose tb. the knowledge base of this system is constructed by taking experts’ suggestions and by analyzing historical data of tb patients. the experiments, carried out, by tak- ing the data of patients demonstrate that the brbes’s this article is part of the topical collection on mobile & wireless health. � karl andersson karl.andersson@ltu.se mohammad shahadat hossain hossain ms@cu.ac.bd faisal ahmed faisalcsecubd@gmail.com fatema-tuj-johora inna.johora@gmail.com department of computer science and engineering, university of chittagong, chittagong, bangladesh pervasive and mobile computing laboratory, luleå university of technology, se- skellefteå, sweden generated results are more reliable than that of human expert as well as fuzzy rule based expert system. keywords expert system · belief rule base · uncertainty · tuberculosis · signs and symptoms introduction tuberculosis (tb) is considered as one of the life threaten- ing infectious diseases all over the world, usually, caused by the bacterium mycobacterium tuberculosis. it is usually two types, namely pulmonary tb (ptb) and extra-pulmonary tb (etb). ptb affects lungs, while etb can attack any organ of the body except brain, spine, heart, pancreas, skele- tal striated muscle, and thyroid. the rate of occurrence of ptb is much higher than that of etb [ – ]. in , about . million people became ill and . mil- lion died from tb all over the world. it has been observed that over % death from tb occurs in low and middle income countries. it is considered as one of the top five causes of death for women aged between to [ ]. the tb bacteria are usually encapsulated as tiny capsules, called tubercles, in the people with healthy immune system. this stage is known as latent tb. in this stage, the bacteria remain inactive and cannot spread to other people. on the contrary, when people’s immune system becomes weak and hence, it is unable to prevent the growth of bacteria. eventually, tb becomes active in the human body. only active pulmonary tb is contagious and the bacteria spread into the air through the cough and sneeze of the affected people. however, etb is not contagious. in case of ptb nearby people can eas- ily be infected during inhaling. tb can be fatal if it is not treated in time, causing serious complications in the lungs, http://crossmark.crossref.org/dialog/?doi= . /s - - - &domain=pdf http://orcid.org/ - - - mailto:karl.andersson@ltu.se mailto:hossain_ms@cu.ac.bd mailto:faisalcsecubd@gmail.com mailto:inna.johora@gmail.com page of j med syst ( ) : forming hole between the nearby airways, making breath- ing difficult because of blocked airways. the primary signs and symptoms of tb are coughing more than three weeks, coughing up blood, fatigue, unintentional weight loss, chest pain, prolonged fever, lack of appetite and night sweating [ – ]. a physician generally determines the suspicion of tb based on these signs and symptoms. signs are measured by physician while symptoms are expressed by the patients [ , ]. patients usually express the symptoms by using lin- guistic terms such as high, medium and low, which are imprecise, ambiguous and vague. therefore, these linguis- tic terms cannot express the level of symptoms with % certainty and hence, it inherits the types of uncertainty mentioned. in some cases, patients may ignore the importance of coughing since they consider it is related to other common diseases, which is an example of uncertainty due to igno- rance. the sputum smear microscopy, which is a method to diagnose the presence of active tb, sometimes it is unable to detect. this is an example of uncertainty due to incompleteness. a comprehensive survey has been car- ried out in consultation with the physicians of the various tb hospitals, located in chittagong district of bangladesh, to identify the types of uncertainties, associated with each of the signs and symptoms of tb, which are described in table . since the traditional way of determining suspicion of tb is usually carried out by the physicians by looking at the signs and symptoms, it does not consider the above uncertain phenomenon. thus, the method jeopardizes the accuracy of the detection of tb. however, an expert sys- tem which emulates the decision making process of human being can be considered as an appropriate tool to address the uncertain phenomenon to accurately detect the suspicion of tb. an expert system consists of two components, namely knowledge-base and the inference mechanisms. however, such an expert system should have the knowledge repre- sentation schema to acquire uncertain clinical knowledge. at the same time, inference mechanism should have the robust reasoning algorithms with the capability to handle various types of uncertainties as mentioned. therefore, in this study the development of a belief rule-based expert sys- tem has been considered, where belief rule base used to handle uncertain knowledge and the evidential reasoning is used as the inference mechanism. the remaining of the article is organized as follows. section “literature review” presents the literature review, “an overview of belief rule based expert system’s method- ology” gives an overview of the belief rule based expert sys- tems methodology, and “brbes to diagnose tb” describes the design, architecture, knowledge-base construction along with an overview of the system interface. section “results and discussion” includes the results and discussion, while “conclusion” concludes the article. literature review several studies have been conducted with reference to the diagnosis of tuberculosis (tb). multilayer neural networks (mlnn) structures were used to facilitate the analysis of tuberculosis [ ]. back-propagation with momentum and levenberg-marquardt algorithms were used to perform the task of training in the mlnn. however, in this approach, an absence of explicit relationship between the input and output data noticed, which is necessary to ensure the trans- parent diagnosis of tb. moreover, this approach has not considered the uncertainty issues, associated with the signs and symptoms, related to the input data. the ensemble classifiers such as bagging, adaboost and random forest tree were used to estimate the performance in detecting pulmonary tuberculosis [ ]. coughing sound detection algorithm [ ] and lung auscultation software [ ] were also used to tb diagnosis. the above approaches con- sidered most of the signs and symptoms as appear in table in diagnosing tb. however, they lack the procedures to measure these signs and symptoms by taking account of the various uncertainties; rather there measuring approach is boolean in nature. in [ ], k-mean clustering was combined with the dif- ferent classifiers, namely, support vector machine (svm), naı̈ve bayesian classifier and k-nearest neighbor (k-nn) to support the prediction of tuberculosis [ ]. moreover, there are systems developed to accurately classify tumor and epilepsy using genetic algorithm by combining multiclass classification method and support vector machine (svm) [ , ]. however, svm lacks the transparency [ , ] since by using this method the relationship between the signs and symptoms of the patient and the diagnosis cannot be estab- lished in an understandable way. however, in case of tb diagnosis, a continuous generation of scenario to establish the relationship between the signs and symptoms and the treatment as well as with the diagnostic process is essential [ , ]. fuzzy cluster means (fcm or fuzzy c-mean) analysis [ , ] used a clustered data set for identifying differ- ent types of tb. tuberdiag [ ] is also a fuzzy rule based system for detecting tb which combines positive and negative knowledge. although fuzzy based expert system can handle uncertainties due to vagueness, imprecision and ambiguity, they are not equipped to address uncertainties due to ignorance, incompleteness and randomness, which are the cases with the signs and symptoms of tb as shown in table . j med syst ( ) : page of table types of uncertainties associated with tuberculosis sign types of uncertainty description coughing ignorance most patients ignore coughing since they consider it is related to a common disease imprecision most patients ignore coughing since they consider it is related to a common disease incompleteness “sputum smear microscopy” is the primary method to diagnose pulmonary tb. however, the method is unable to detect half of all active tb cases coughing up blood ignorance sometimes blood remains unnoticed due to thick and larger amount of phlegm fatigue imprecision, vagueness, ambiguity people consider fatigue as a normal symptom of overwork- ing and do not consult with doctor prolonged fever ignorance, vagueness, randomness usually patients can not exactly tell the duration and temperature when doctor asks about the fever night sweating ignorance, vagueness most people think that warm weather, humidity or wear- ing heavy cloths are liable for night sweating and hide the symptom from the doctor in addition, the above mentioned approaches [ – , – ] lack the appropriate inference procedures to address different types of uncertainties such as vagueness, impreci- sion, ambiguity, ignorance, incompleteness and randomness in a single integrated framework. such an integrated frame- work has an important role to make robust decision to support tb diagnostic decision making process. therefore, it is necessary to employ an appropriate knowledge representation schema to capture uncertain knowledge that exists with the signs and symptoms of tb. belief rule base (brb) is widely used to represent this type of knowledge [ – ]. in addition, brb can be used to demonstrate the explicit non-linear relationship between the input-output data which is necessary to ensure the trans- parent diagnosis of tb. evidential reasoning in combination with the brb can be used as the inference mechanism of the expert system which has the capability to handle all types of uncertainties in an integrated framework [ – ]. there- fore, the following section will represent the belief rule based expert systems (brbess) methodology, consisting of knowledge-base construction and the inference procedures. an overview of belief rule based expert system’s methodology an expert system is mainly consists of two components, namely, knowledge-base and the inference procedures. in brbess, belief rule-base is used to represent the domain knowledge under uncertainty. on the other hand, inference procedures of brbes consists of input transformation, rule activation weight calculation, belief update and rule aggre- gation using evidential reasoning [ ]. each of them will be elaborated below. domain knowledge representation using brb a belief rule is the extension of traditional if-then rule, where a belief structure is used in the consequent part. antecedent part of the belief rule consists of one or more antecedent attributes with associated referential values, while consequent part consists of one consequent attribute. knowledge representation parameters such as rule weight and antecedent attribute weight are used. a belief rule base (brb) consists of one or more than one belief rules. the reason for adopting if-then rule is that it is considered as an appropriate mechanism to represent human knowledge [ ]. in addition, non-linear causal relationship between the antecedent and consequent attributes can be established in a belief-rule-base. equation represents an example of belief rule. rk : ⎧ ⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎨ ⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎩ if (a is a k ) ∧ (a is ak ) ∧ ... ∧ (atk is aktk ) then ⎧ ⎪⎪⎪⎪⎪⎪⎨ ⎪⎪⎪⎪⎪⎪⎩ {(c , βk ), (c , βk ), ..., (cn , βkn )}, (( n∑ n= βkn ≤ )), with rule weight ≤ θk ≤ , and attribute weight δk , δ k , ..., δ k t ≥ satisfying tk∑ i= δk i = ( ) where, a , a , ..., a k tk , tk ∈ { , , ..., t }, are the antecedent attributes used in the kth rule. ak i ∈ {ai , ai , ..., aiji } is the referential value of antecedent attribute. c , c , ..., cn are the referential values of the con- sequent attribute while βki is the belief degree to which ci page of j med syst ( ) : is believed to be true. if n∑ i= βki = , the belief rule is said to be complete, otherwise it is incomplete. in this way, brb addresses the uncertainty resulting from the incompleteness. equation represents the example of a belief rule from the domain of tb. here, the consequent attribute is “tb suspicion” with three referential values consisting of “high”, “medium”, and “low” with degree of belief . , . , and . . the rule is said to be complete since the summation of the belief degrees stands at . ⎧ ⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎨ ⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎩ if ⎧ ⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎨ ⎪⎪⎪⎪⎪⎪⎪⎪⎪⎪⎩ (coughing more than weeks is medium) and (coughing up blood is high) and (chest pain is high) and (fatigue is low) and (prolonged fever is medium) and (lack of appetite is low) and (weight loss is high) and (night sweating is low) then tb suspicion is (high, . ), (medium, . ), (low, . ) ( ) each antecedent attribute of this rule also consists of three referential values and hence, the total number of rules in this brb can be calculated by applying eq. , which is , . l = t∏ i= (ji ) ( ) where l is the total number of rules in a sub rule base and ji is the referential values of the ith antecedent attribute. brbess inference procedures each of the inference procedures used in a brbes is described below. input transformation input transformation consists of the distribution of the input value of an antecedent attribute over its different referential values by applying eq. . the distributed value with each referential values of the antecedent attribute is called match- ing degree or the degree of belief. it is interesting to note that when the value of this matching degree is calculated for each of the referential values of the antecedent attribute (aa), this value is assigned to only those rules where this referential value exits with the aa. as an example, eq. consists of eight antecedent attributes, each with three referential values. when the input data for one antecedent attribute is collected from the tb patient, its matching degrees to the corresponding referen- tial values are calculated by applying eq. . consequently, the matching degree, related to a referential value corre- sponding to the antecedent attribute of a rule is assigned. in this way, input data for the eight antecedent attributes can be collected for a patient and their corresponding match- ing degree is assigned in a rule. once the rule is assigned with the corresponding matching degree then it is said to be active and the rule is called packet antecedent. this phe- nomenon can be described that the rule is in the ram while the initial rule base is in the secondary memory. αih = u(aih+ ) − a∗i u(aih+ ) − u(aih) and αih+ = a∗ i − u(aih) u(aih+ ) − u(aih) ( ) where u(aih+ ) and u(aih) are grade values of aih+ and aih respectively. table shows the matching degree of the antecedent attribute value into its different referential values. for exam- ple, the input value of the antecedent attribute ”cough” is identified as ”low”, which is weighted as % by the expert and its corresponding matching degrees associated with the referential values (in this case they are “high”, “medium” and “low”) obtained by using eq. . calculation of activation weights rule activation weight calculation consists of calculating the combined matching degree (αk ) as well as the weight of a rule in the brb. equation consists of eight antecedent attributes and hence, it is important to calculate their com- bined matching degree, which can be calculated by using multiplicative function [as shown in eq. ], allowing the inter-relationship between the attributes [ ]. αk = tk∏ i= (α k i ) δ k i , δ k i = δk i max i= ,...,tk (δ k i ) ( ) so that ≤ δk i ≤ where δk i (i = , ..., tk) is the rela- tive weight of the ith antecedent attribute in the kth belief rule. tk is the total number of antecedent attributes in the kth rule. here, δk i = is meaning that the attribute which has zero importance and hence, no impact on the aggre- gation process, while δk i = demonstrates the significant impact. moreover, overall belief increases with the incre- ment of the individual belief of the antecedent attributes. the rule activation weight is calculated by using eq. [ ]. wk = {θkαk} { l∑ i= θi αi } ( ) j med syst ( ) : page of where θk is the relative weight of the kth rule and l is the total number of belief rule in the belief rule-base. when the wk of a rule is zero then it has no impact in the brb while it is “ ” then its important is high. belief degree update equation shows the presence of eight antecedent attributes, necessary to assess the suspicion of tb. however, there could be some situation when the data of the some attributes could not be available. in such situation, the initial belief degrees that were assigned to the referential values of the consequent attribute needs to be updated by applying eq. [ ]. this phenomenon is an example uncertainty due to ignorance. βki = βki ∑t k t= (τ (k, t) ∑ji j = αtj ) ∑tk t= τ (k, t) ( ) where τ (k, t) = ⎧ ⎨ ⎩ , if used in defining rk(t = , ..., tk) or , otherwise here, βki is the original belief degree while βki is the updated belief degree. the original belief degree is updated while any ignorance is noticed. for example, if the antecedent “cough” is ignored, then the initial belief degrees are updated as shown in table . inference using evidential reasoning in order to obtain the aggregated value of the referential val- ues of the consequent attribute, based on the input data of the antecedent attributes, either recursive or analytical evi- dential reasoning (er) algorithms as shown in eq. can be applied [ ]. to reduce the computational complexity ana- lytical er algorithm is found effective to calculate the final belief degree βj . βj = μ × [∏lk= (ωkβkj + − ωk ∑n j = βkj ) − ∏l k= ( − ωk ∑n j = βkj )] − μ × [∏lk= ( − ωk)] ( ) with μ = [∑nj = ∏l k= (ωkβkj + − ωk ∑n j = βkj ) − (n − ) × ∏lk= ( − ωk ∑n j = βkj )]− . the final combined result or output generated by er is represented by {(c , β ), (c , β ), ..., (cn , βn )} where βj is the final belief degree attached to the j th referential value cj of the consequent attribute c, which is obtained after all activated rules in the brb are combined by using er. this output can be converted into a crisp/numerical value [as shown in eq. ] by assigning a utility score to each referential value of the consequent attributes [ ]. h (a ∗ ) = n∑ j = u(cj )βj ( ) where h (a∗) is the expected score expressed as a numerical value and u(cj ) is the utility score of the j th referential value. brbes to diagnose tb this section presents the architecture along with imple- mentation strategy of the belief rule-based expert system (brbes) to diagnose tuberculosis (tb). this is followed by the presentation of the knowledge-base construction as well as a description of the brbes’s interface. architecture and implementation a system architecture can be defined how its compo- nents are organized. it is also important to know the pattern of system organization, which is known as archi- tectural style. brbes presented in this article follows three-layer architecture, consisting of web-based interface, application and database management layers as shown in fig. . web-based interface: since the brbes to be used by the physicians, patients and researchers at various hospitals of bangladesh, especially in the rural areas, a web-based user-friendly interface is necessary. therefore, to ensure the usability of the system by the mentioned users at any time and at any place, a web-based interface has been devel- oped for the brbes. this interface has been developed by integrating various web technologies such as javascript, jquery, html and css. the application layer, which facil- ities inference and database access, has been developed by using php because of its simplicity, shorter develop- ment cycle and it can be used through online. the data- base management layer, which consists of clinical data and knowledge-base, developed by using mysql, which is a relational database management system. mysql is exible, user friendly and ensures security as well as faster data access. page of j med syst ( ) : table input transformation serial no. antecedent name antecedent value high medium low cough low, % . . blood with cough high, % . . chest pain high, % . . fatigue low, % . . fever high, % . . lack of appetite medium, % weight loss high, % . . night sweating low, % . . knowledge base construction in brb the knowledge-base construction consists of developing a brb tree by identifying the necessary antecedent and con- sequent attributes. figure shows the single level brb structure to diagnose tb. the leaf nodes represent the antecedent attributes while root node consequent attribute. the eight antecedent attributes, having three referential val- ues each, are identified and they are verified in consultation with the physicians, located at the various hospitals of chittagong city of bangladesh. the brb consists of , rules since it comprises eight antecedent attributes each with three referential values. the number of rules are calculated by using eq. . a brb usu- ally can be established [ ] by acquiring domain expert knowledge, by collecting historical data, by using earlier rules if they are available and by developing rules in a random way without any prior-knowledge. in this study, the initial brb has been constructed by acquiring knowledge of physicians or the domain experts as well as by using patients’ historical data. each rule of the brb has given rule weight “ ” while each antecedent attribute’s weight is considered as “ ”. an example of such rule is illustrated in eq. . table illustrates the initial brb of the brbes. brb interface an interface can be defined as a media, facilitating the users to interact with the system. figure illustrates the interface of the brbes to diagnose tb, allowing the acquisition of the input value, associated with each of the eight antecedent table belief degree update rule id high medium low initial . . update . . fig. architecture of brbes fig. brb framework to assess tb j med syst ( ) : page of table a sample of initial belief rule-base for assessment of tuberculosis suspicion rule no rule weight antecedents consequent if then a a a a a a a a tb suspicion high medium low h h h h l h h h . . . h l m l h l l l . . . l h l m h l l m . . . l l l l h l l l . . . h m h l h l l h . . . m h h l m l h l . . . h m m m h l l m . . . m l l l h m m l . . . h h h h h h h m . . . l m h h h m m l . . . attributes, either from the patients or from the physicians. figure illustrates the matching degree, associated with the antecedent attributes with reference to the data of the first patient as shown in table . for example, the input value of the antecedent attribute a (coughing more than three weeks) is obtained as “ ”. this is acquired by asking the patient about the intensity level of the a , which is in the range of - . this input value is then distributed over the three referential values of a , which are “(high, . )”, “(medium, . )” and “(low, . )”. this distributed value is called the matching degree obtained by using eq. . the suspicion level of the tb of this patient is obtained by using eqs. , , and . this is measured as degree of belief associated with each of the referential values of the suspicion level of tuberculosis and found as (high, . ), (medium, . ), and (low, . ). these fuzzy values are converted into crisp value by using eq. and obtained as . %. fig. brbes interface to assess tb page of j med syst ( ) : table tb suspicision by fuzzy logic, brbes, and expert sl. no ( ) a ( ) a ( ) a ( ) a ( ) a ( ) a ( ) a ( ) a ( ) fuzzy result(%) ( ) brbes result(%) ( ) expert opinion(%) ( ) bench- mark ( ) . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . results and discussion to demonstrate the applicability and the reliability of the brbes to diagnose tuberculosis (tb), the system fed with the input data received from the tb patients of a hospital located in the chittagong city of bangladesh (fig. ). the input data, associated with the eight attributes, of tb patients have been collected. the real laboratory test results of those patients collected and they were con- sidered as the benchmark data. if tb is found positive in laboratory test result for a patient, then the benchmark is considered as “ ”, otherwise it is “ ” for that patient. table. v illustrates the collected data on the eight attributes of a patient (column through ) along with level of tb suspi- cion generated by the brbes (column ). table. v also illustrates the expert opinion on the level of tb suspicion fig. tuberculosis hospital, chittagong, bangladesh j med syst ( ) : page of fig. roc curves comparing the result of brbes and expert data (column ) and the benchmark data is recorded in col- umn . in order to compare the reliability of the brbes’s results a fuzzy rule based expert system (frbes) to measure the suspicion level of tb developed in matlab environment and the results generated for the same data by using frbes are recorded in column . for sim- plicity, table only presents the data of patients out of . fig. roc curves comparing the result of brbes, frbes and expert data page of j med syst ( ) : the receiver operating characteristic (roc) curves are usually used to analyze the accuracy and reliability of the diagnostic tests having ordinal or continuous results. there- fore, the method was considered, to measure the reliability of brbes in comparison with expert opinion and frbes. the accuracy of a system to assess the level of suspicion of the tb can be measured by calculating the area under curve (auc) [ , ]. for example, if the auc is found to be one for the results generated by the brbes then the system can be considered as % reliable. figure illus- trates the roc curves plotted for both the brbes and the expert opinion. the roc curve plotted by the blue line in this figure is associated with the results generated by the brbes with auc of . ( % confidence intervals . - . ). the roc curve plotted by the green line in fig. is obtained against the physician’s opinion, and its auc is . ( % confidence intervals . - . ). however, fig. illustrates the roc curves for brbes, frbes and expert opinion. the roc curve plotted by the red line in fig. is obtained against the frbes and its auc is . ( % confidence intervals . - . ). while the auc values of fig. for both the brbes and the expert opinion are same as of fig. . table summarizes the above results associated with brbes, frbes and expert opinion. from figs. and as well as from table it can be observed that auc of expert opinion is much less than from both the brbes and frbes. the reason for this is that dur- ing the interviewing and conversation with the physicians it has been observed that they are not aware of the uncertainty issues related to the signs and symptoms of the tb. the reason for this is that during the interviewing and conversation with the physicians it has been observed that they are not aware of the uncertainty issues related to the signs and symptoms of the tb. therefore, the reliability of their assessment level of tb suspicion is much lower than that of brbes and frbes. from the table as well as from figs. and , it can also be observed that auc of brbes is much larger than that of frbes. the reason for this is that fuzzy rule based expert system (frbes) only considers uncertainty due to the vagueness, imprecision and ambiguity. however, brbes includes uncertainties due to table reliability comparison among three systems test result variables area asymptotic % confidence interval lower bound upper bound brbes . . . fuzzy system . . . expert data . . . the ignorance, incompleteness and randomness in addition to the vagueness, imprecision and ambiguity. further, the inference procedures of the frbes which uses either mamdanior t-s methods are not equipped to pro- cess uncertainty issues during the reasoning process. on the contrary, brbes uses evidential reasoning procedures as the inference engine which is equipped to handle types of uncertainties mentioned before. conclusion this article described the design, implementation and appli- cations of a belief rule based expert system (brbes), allowing the measurement of the level of suspicion of tb by taking account of its various signs and symptoms. the system allows the understanding of the relationship between signs and symptoms and the level of suspicion of tb of a patient in an explicit and transparent way. this will allow the identification of the signs and symptoms those are respon- sible for increasing the suspicion level of tb of a patient. in this way, various scenarios by taking account of the signs and symptoms of a patient from various perspectives can be carried out by using the brbes. the physicians will eventually be able to select the appropriate medicine for a patient. in addition, all types of uncertainties such as vague- ness, imprecision, ambiguity, ignorance and incompleteness can be handled in an integrated framework which makes the system more robust as evident from the comparative results generated by using both the manual system and the fuzzy rule based expert system as illustrated in table . there- fore, the brbes can provide a decision support platform to the physicians and would serve a savior by offering primary health care to the people with reduced time and low diagno- sis cost. in future, the bebes would be extended to support optimal learning by training the knowledge representation parameters such as rule weight, attribute weight, and belief degrees. open access this article is distributed under the terms of the creative commons attribution . international license (http:// creativecommons.org/licenses/by/ . /), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the creative commons license, and indicate if changes were made. compliance with ethical standards conflict of interests author mohammad shahadat hossain declares that he has no conflict of interest. author faisal ahmed declares that he has no conflict of interest. author fatema-tuj-johora declares that she has no conflict of interest. author karl andersson declares that he has no conflict of interest. http://creativecommons.org/licenses/by/ . / http://creativecommons.org/licenses/by/ . / j med syst ( ) : page of funding this study was funded by swedish research council under grant - . ethical approval all procedures performed in studies involving human participants were in accordance with the ethical standards of the institutional and/or national research committee and with the helsinki declaration and its later amendments or comparable ethical standards. informed consent informed consent was obtained from all individ- ual participants included in the study. references . walker, b.r., colledge, n.r., ralston, s.h., and penman, i.d. davidson’s principles and practice of medicine. nd ed. united kingdom: elsevier health sciences, . . goldman, l., and bennet, j.c. cecil textbook of medicine. th ed. philadelphia: elsevier saunders, . . walker, h.l., chen-yuan, c., gie, r.p., and enarson, d.a. crofton’s clinical tuberculosis. rd ed. london: macmillan edu- cation, . . whotuberculosis, http://www.who.int/mediacentre/factsheets/ fs /en/, accessed: january , . . karim, r., hossain, m.s., khalid, m.s., mustafa, r., and bhuiyan, t.a.: a belief rule based expert system to assess bronchiolitis suspicion from signs and symptoms under uncertainty. in: sai intelligent systems conference (intellisys). london, uk, . . hossain, m.s., zander, p.-o., kamal, m.s., and chowdhury, l., belief-rule-based expert systems for evaluation of e- government: a case study. expert. syst. ( ): – , . . er, o., temurtas, f., and çetin tannkulu, a., tuberculosis disease diagnosis using artificial neural networks. j. med. syst. ( ): – , . . rusdah, e., and wardoyo, r.: winarko preliminary diagnosis of pulmonary tuberculosis using ensemble method. in: interna- tional conference on data and software engineering (icodse). yogyakarta, indonesia, . . tracey, b.h., comina, g., larson, s., bravard, m., lópez, j.w., and gilman, r.h.: cough detection algorithm for monitoring patient recovery from pulmonary tuberculosis. in: annual international conference of the ieee engineering in medicine and biology society. boston, ma, usa: embc, . . lestari, r., ahmad, m., alisjahbana, b., and djatmiko, t.: the lung diseases diagnosis software: influenza and tuberculosis case studies in the cloud computing environment. in: international conference on cloud computing and social network- ing (icccsn). bandung, indonesia, . . asha, t., natarajan, s., and murthy, k.n.b.: a data mining approach to the diagnosis of tuberculosis by cascading clustering and classification. j. comput. ( ), . . shen, c.p., lin, j.w., lin, f.s., lam, a.y.y., chen, w., zhou, w., sung, h.y., kao, y.h., chiu, m.j., leu, f.y., and lai, f.: ga-svm modeling of multiclass seizure detector in epilepsy analysis system using cloud computing. soft. comput. – , . doi: . /s - - - . . lin, f.s., shen, c.p., liu, c.h., lin, h., huang, c.y.f., kao, c.y., lai, f., and lin, j.w., a high performance multiclass classifica- tion framework using cloud computing architecture. journal of medical and biological engineering ( ): – , . . calzada, a., liu, j., wang, h., nugent, c., and martinez, l., application of a spatial intelligent decision system on self-rated health status estimation. j. med. syst. ( ): – , . . huysmans, j., dejaeger, k., mues, c., vanthienen, j., and baesens, b., an empirical evaluation of the comprehensibility of decision table, tree and rule based predictive models. decis. support. syst. ( ): – , . . ainon, r.n., bulgiba, a.m., and lahsasna, a., ami screening using linguistic fuzzy rules. j. med. syst. ( ): – , . . bojarczuk, c.c., lopes, h.s., and freitas, a.a., genetic program- ming for knowledge discovery in chest pain diagnosis. ieee eng. med. biol. mag. ( ): – , . . soundararajan, k., sureshkumar, s., and anusuya, c., diagnos- tics decision support system for tuberculosis using fuzzy logic. international journal of computer science and information tech- nology & security (ijcsits) ( ): – , . . imianvan, a.a., and obi, j.: fuzzy cluster means expert system for the diagnosis of tuberculosis. global j. comp. sci. technol. ( ), . . phuong, n.h., hung, d.h., tuan, d.t., co, n.v., duong, b.d., thinh, p.t., hoa, n.p., and linh, n.n.: designing an experimen- tal expert system for lung tuberculosis diagnostics using fuzzy set theory. in: ieee international conference on systems, man, and cybernetics. san diego, ca, usa, . . yang, j.-b., liu, j., wang, j., sii, h.-s., and wang, h.-w., belief rule-base inference methodology using the evidential rea- soning approach-rimer. ieee transactions on systems man, and cybernetics-part a: systems and humans ( ): – , . . hossain, m.s., monrat, a.a., hasan, m., karim, r., buhiyan, a.t., and khalid, m.s.: a belief rule based expert system to assess mental disorder under uncertainty. in: th interna- tional conference on informatics, electronics & vision (iciev). dhaka, bangladesh, . . hossain, m.s., andersson, k., and naznin, s.: a belief rule based expert system to diagnose measles under uncertainty. in: proceed- ings of the international conference on health informatics and medical systems (hims ). las vegas, nevada, usa, . . hossain, m.s., hasan, m.a., uddin, m., islam, m.m., and mustafa, r.: a belief rule based expert system to assess lung cancer under uncertainty. in: th international confer- ence on computer and information technology (iccit). dhaka, bangladesh, . . karim, r., andersson, k., hossain, m.s., uddin, m.j., and meah, m.p.: a belief rule based expert system to assess clinical bron- chopneumonia suspicion. in: future technologies conference (ftc ). san francisco, ca, usa, . . hossain, m.s., haque, m.a., mustafa, r., karim, r., dey, h.r., and yousuf, m.: an expert system to assist the diagnosis of ischemic heart disease. int j integr care, . . hossain, m.s., khalid, m.s., akter, s., and dey, s.: a belief rule-based expert system to diagnose influenza. in: th international forum on strategic technology (ifost). cox’s bazar, bangladesh, . . rahaman, s., and hossain, m.s.: a belief rule based clinical deci- sion support system to assess suspicion of heart failure from signs, symptoms and risk factors. in: international conference on informatics, electronics & vision (iciev). dhaka, bangladesh, . . metz, c.e., basic principles of roc analysis. semin. nucl. med. ( ): – , . . hanley, j.a., the robustness of the binormal assumptions used in fitting roc curves. med. decis. mak. ( ): – , . http://www.who.int/mediacentre/factsheets/fs /en/ http://www.who.int/mediacentre/factsheets/fs /en/ http://dx.doi.org/ . /s - - - a belief rule based expert system to assess tuberculosis under uncertainty abstract introduction literature review an overview of belief rule based expert system's methodology domain knowledge representation using brb brbess inference procedures input transformation calculation of activation weights belief degree update inference using evidential reasoning brbes to diagnose tb architecture and implementation*. pt knowledge base construction in brb brb interface results and discussion conclusion open access compliance with ethical standards conflict of interests funding ethical approval informed consent references this article appeared in a journal published by elsevier. the attached copy is furnished to the author for internal non-commercial research and education use, including for instruction at the authors institution and sharing with colleagues. other uses, including reproduction and distribution, or selling or licensing copies, or posting to personal, institutional or third party websites are prohibited. in most cases authors are permitted to post their version of the article (e.g. in word or tex form) to their personal website or institutional repository. authors requiring further information regarding elsevier’s archiving and manuscript policies are encouraged to visit: http://www.elsevier.com/copyright http://www.elsevier.com/copyright author's personal copy modelling situation awareness for context-aware decision support yu-hong feng, teck-hou teng, ah-hwee tan * intelligent systems centre and school of computer engineering, nanyang technological university, nanyang avenue, singapore , singapore abstract situation awareness modelling is popularly used in the command and control domain for situation assessment and decision support. however, situation models in real-world applications are typically complex and not easy to use. this paper presents a context-aware decision support (cads) system, which consists of a situation model for shared situation awareness modelling and a group of entity agents, one for each individual user, for focused and customized decision support. by incorporating a rule-based inference engine, the entity agents provide functions including event classification, action recommendation, and proactive decision making. the implemen- tation and the performance of the proposed system are demonstrated through a case study on a simulated command and control application. � elsevier ltd. all rights reserved. keywords: situation awareness; context awareness; decision support; intelligent agents . introduction the concept of situation awareness is well established in the field of human factor studies in complex environments. according to endsley ( ), situation awareness refers to ‘‘the perception of the elements in the environment within a volume of time and space, the comprehension of their mean- ing and the projection of their status in the near future’’. in the domain of command and control, wherein a designated commander is required to exercise authority and direction over various forces in order to achieve the given goals, a clear picture of the current situation and an accurate pro- jection of the future states are essential for effective deci- sion making. as such, situation awareness modelling has become an important component of command and control decision support systems. over the years, guidelines and processes have been developed for building situation awareness models from a goal-oriented perspective (ends- ley, ). however, to the best of our knowledge, none has incorporated the notion of context awareness for pro- viding customized situation awareness. context awareness was introduced by schilit and thei- mer ( ) to develop software that adapts according to its locations of use, the collection of nearby people and objects, as well as changes to those objects over time. with technology advancements and the growing interest in mobile and wearable computation devices, context aware- ness has become one of the major research areas in those fields. the definition of context has also expanded from physical attributes to include device characteristics as well as user-specific factors, such as profiles and preferences (dey, ; harter, hopper, steggles, ward, & webster, ; moran & dourish, ). whereas situation awareness focuses on the modelling of a user’s environment so as to help the user to be ‘‘aware of his current situation’’, context awareness is about exploiting the context of a user and helping the user to have a more effective interaction with the system by actively changing the system’s behavior according to the user’s cur- rent context or situation. in the domain of command and control, individual users require specific sets of situation awareness. also, the same piece of information may have different meanings and usages for different people in the - /$ - see front matter � elsevier ltd. all rights reserved. doi: . /j.eswa. . . * corresponding author. e-mail addresses: yhfeng@ntu.edu.sg (y.-h. feng), teng @ntu. edu.sg (t.-h. teng), asahtan@ntu.edu.sg (a.-h. tan). www.elsevier.com/locate/eswa available online at www.sciencedirect.com expert systems with applications ( ) – expert systems with applications author's personal copy same environment. therefore, the system must know the context of the current user and forward focused contextual information to the user. building a single situation model and presenting the full set of information to all users is not only inefficient but highly confusing. on the other extreme, building separate situational awareness models for individual users is not only inefficient but more impor- tantly, raises the issue of consistency. in this paper, we present a system known as cads (for context-aware decision support) that incorporates a shared situation awareness model but provides individual human operators with customized views and services through a group of entity agents. the entity agents, one for each individual user, communicate with the situation model and extract information of relevance for presenta- tion to their respective users in accordance to the user con- text. to reduce the cognitive load of the human operators, the entity agents perform event classification and action recommendation in a proactive manner. we have applied cads to a simulated command and control domain and evaluated its performance based on several mission scenar- ios. our experimental results show that the entity agents are able to perform event classification and action recom- mendation with a high level of accuracy and thereby reduce the cognitive load of the human operators significantly. the rest of the paper is organized as follows. section presents the overall system architecture. section discusses the various components in the shared situation awareness model. section presents the design and implementation of the entity agents for context-aware decision support. section illustrates the various system functionalities through a case study on the command and control applica- tion. section describes our evaluation methodology and reports the experimental results. section discusses related work. the final section concludes and outlines the future work. . system overview referring to fig. , the cads architecture incorporates a situation model managing the shared situation awareness model and a group of entity agents, supported by an under- lying game simulation engine. the game simulation engine gecco (for game environment for command and con- trol operations) gecco ( ) is a publicly available gen- eric platform for creating and playing real-time multi- player strategy games. the situation awareness model is based on the endsley’s three layer structure (endsley, ), but with the addition of a terrain model for the command and control domain. assuming the existence of a sensor network, incoming data are first represented at the perception layer (level ) of the situation model. the comprehension layer (level ) then interprets the data and provides assessment of the current situation. to support anticipatory decision making, the projection layer (level ) predicts future states based on the understanding of the current situation. at present, only some basic projection functions have been implemented for illustration purpose. all three layers of the situation aware- ness model function concurrently and iteratively. each entity in the environment is represented by an entity agent. each agent has a set of goals and strategies, which form the basis of the entity’s behaviors. the goals and strategies are in turn translated into executable struc- tures like missions, plans and actions. together with the physical attributes such as location and strengths, all these constitute the context based on which an agent provides a customized set of views and services of the shared situation awareness model to its user. given a user context, an entity agent pro-actively scans the situation awareness model for relevant information in the environment. it then highlights important events to the user according to their significance in the given context. our current implementation of the entity agents is based on the production rule representation and a logical reasoning mechanism. . situation awareness modeling we describe the various levels of the situation awareness model, including the terrain model, in the following sections. . . terrain model representing the knowledge of the physical world, ter- rain information forms the basis of all three layers of situ- ation awareness as well as context-aware decision support. as such, the situation model must be able to monitor the dynamic changes in the terrain, which in turn may result in significant events in the situation awareness model. the terrain model consists of two key elements, namely nodes and links. a node represents a critical point or loca- tion on the map with the necessary descriptive attributes. a link corresponds to a connection or path between a given pair of nodes, with the necessary attributes describing the condition of the connection as well as a weight attribute providing distance information. the terrain model further incorporates a shortest path algorithm for identifying the optimal route between any two nodes in the terrain. to facilitate the process of building and editing the ter- rain model, an interactive software tool named terrainfig. . the cads system architecture. y.-h. feng et al. / expert systems with applications ( ) – author's personal copy builder (fig. ) has been developed in-house. using the notions of nodes and links, a terrain model can be easily constructed from a two-dimensional map by clicking on the appropriate positions on the map. . . sa level : perception the key elements in the level of the situation model are entities and events. an entity represents an object in a sit- uation which has attributes such as identities, capabilities, and so on. virtually everything in the environment can be an entity. as shown in fig. , the entity class is a general description of an object in a situation with the basic and necessary attributes, such as identity, trajectory, and child- entities. the trajectory is a sequence of movement made by the entity. the event class is a data structure encapsulating all the relevant information of a physical occurrence in a situa- tion. these informations are decomposed into five ele- ments, i.e. the five w’s, namely when, where, who, what and why. an event injection triggers the system to reassess the relevant entities’ attributes and their relations with the others. these changes in turn result in a new state of situation awareness and may eventually lead to critical decisions. . . sa level : comprehension the comprehension level makes sense out of the data provided by the perception level and integrates them into meaningful pieces of information. consider a movement scenario as an example, in which a space vehicle is tasked to move from one location to another and to avoid aliens along the way. in this scenario, a list of high level functions is essential for providing situation comprehension for indi- vidual entities. an example list of such functions is given below. • distance: calculating the distance from the current loca- tion to the destination location. • speed required: calculating the speed required to reach the destination on time. • fuel required: calculating the fuel required to reach the destination based on the current consumption rate. • enemy on the route: scanning the planned route for enemy. • enemy around the route: scanning the areas around the planned route for potential threats.fig. . the terrain builder. entity conceptualentityphysicalentity -position : [double, double] -condition : string -volitional : boolean movableentity -capabilities : string -speed : int -fuel : int -identity : string -trajectory : trajectory -childentities : entity event -when : string -where : string -who : string -what : string -why : string relation -relationtype : string …* …* …* …* fig. . the schema of the situation model. y.-h. feng et al. / expert systems with applications ( ) – author's personal copy • enemy behind: checking whether there is any enemy coming up from behind. • zone of alert: evaluating the conceptual proximity of enemy in terms of zones. by exploiting the relative location information provided by the enemy detection functions and the proximity informa- tion by the zone of alert functions, an entity is in a better position to formulate its strategies and plans. for instance, if an enemy is on the planned route and in zone (meaning far away), an entity might choose to continue on this route as the enemy is still relatively far away and it may eventu- ally go out of its route. however, if the enemy is in zone (meaning near), the user might choose to re-plan his route. . . sa level : projection the ability to foresee the future is especially crucial in command and control. based on the understanding of the current situation, it is possible to predict the future states to a certain extent. for illustration, we have imple- mented three types of projection functions: enemy location projection, terrain projection and goal status projection. enemy location projection is responsible for predicting the future locations and the corresponding timing of a hos- tile entity. the system keeps a finite history of the time– space information on the enemy. statistical inference is then performed over these historical data in order to esti- mate the entity’s expected location at a particular time. data points far back in the history are either ignored or given a lower priority in estimating the new locations. whenever there is a change in terrain data, terrain pro- jection is conducted to predict the possible future changes, based on the existing knowledge of terrain changes. for example, if a road is blocked due to a heavy rain and our prior knowledge informs that it usually takes up to three hours for a blocked road to clear, the system would then expect to observe another terrain change to unblock the road within three hours. as mentioned, an important agenda of an entity is goal attainment. thus an entity needs to assess the goal status at a regular time interval. using the present information, such as the planned route, the entity’s fuel level and its current speed, the projection layer of the situation awareness model is able to compute the expected time to accomplish the goals. all these projected situations then form a part of the situation awareness. . context-aware decision support the primary role of an entity agent is to provide context- aware decision support in a command and control setting so that decisions can be made in a timely and effective manner. to this end, we have developed entity agents with a range of context-aware capabilities, including event classification, action recommendation and decision mode selection. refer- ring to fig. , each entity agent communicates with the situation model and employs a rule-based inference engine to provide decision support to its user according to the user context model. the user context used by the agents consists of goals, plans and physical attributes, such as location and capabilities. an agent’s goal defines the target states of the corresponding entity. an agent’s plan contains the planned sequences of actions to achieve these goal states. . . the inference engine as the activities of an entity should be driven by its goals, its attention must be directed accordingly. using an iterative process, the attainment of goals must be con- tinuously monitored and assessed. based on the user con- text, especially the goals and plans, an entity agent should thus constantly look out for situations in the envi- ronment that may affect its goal attainment. our current implementation of the entity agents is based on a production rule based engine known as drool (rupp, ), which is publicly available and fully open- source. being one of the established ways to implement sit- uation awareness, production rules provide us with a sim- ple and comprehensible specification for recording the heuristics articulated by the domain experts. based on the forgy’s rete algorithm, the rules in drools can be written in java, python and groovy via the use of xml. each rule in drools consists of a list of parameters, the if conditions and the then consequence. in addition, a salience attribute can be assigned to each rule, as a form of priority in the activation queue. the working memory is stateless and it contains all the knowledge or facts. facts can be asserted, modified and retracted from the working memory. rules have to be loaded into the working memory before firing. a rule firing (execution) can modify the con- tent of the working memory and therefore possibly trigger other rules to fire. . . event classification to reduce the cognitive load of human operators, infor- mation should be presented to a user only if it is of rele- vance and significance to the user. in our system, a user corresponds to an entity in the simulated world. for exam- ple, an explorer commander is represented as an explorer fig. . the entity agent model. y.-h. feng et al. / expert systems with applications ( ) – author's personal copy entity in the simulation. therefore, the explorer com- mander should only receive relevant and significant infor- mation with respect to his/her context. specifically, information should be presented according to their impact on the goals of an entity. depending on how an event may affect the goal attainment status of the entity, it is presented to the user in one of the three message levels: events that do not affect goals are posted as infor- mation, events that may affect goals are issued as warning, and events that endanger goals are flagged as alert. the event classification rule module assigns an appropri- ate message level to each incoming event that would be pre- sented to the user according to the situation assessed. a typical event classification rule for an explorer is as shown below. the terrainrouteblocked rule checks that if the sit- uation type is terrain and the event will affect the current route of the entity, the message level is set to alert. rule name = terrainrouteblocked if situationtype==terrain affectcurrentpath is true then msglevel = alert . . option generation and evaluation for effective decision making in a given situation, a human operator needs to gain a good level of situation awareness by assessing his current situation. with the situ- ation awareness, he can then consider the options of actions that can be performed and decide on the best options available. an entity agent facilitates this process by monitoring the current situation and recommending possible courses of actions to its user when the need arises. the generate choice rule module is responsible for deduc- ing choices or options for a particular situation. currently, there is a total of five action choices available for any given situation, namely resume, increase speed, decrease speed, reroute and wait. the rules determine if a choice is applica- ble in a given situation and generate a list of appropriate choices from the set of actions. for instance, the following rule illustrates the generation of a choice. under the situation type of terrain, if the rule detects that the current route is blocked and an alternate route is available, a reroute choice is created and inserted into the choice list. rule name = situationt if situationtype==terrain currentpathisblocked is true alternaterouteavailable is true then action = reroute given the current situation, the entity agent goes through the complete set of rules, each of which checks for specific conditions and generates a choice when appro- priate. the result of this process is a list of action choices ready for ranking. the rank choice rule module assesses each of the avail- able choices for a given situation with a score between and . generally, the scores are calculated as a function of minimal changes required and payoff in terms of goal attainment. thus, the option with the minimal change is preferred among those with the same payoff. for example, in the movement scenarios, a reroute action is a high change option, whereas resume is a low change action. therefore, given the same level of payoff, resume is pre- ferred over reroute. . . mode selection to exploit their proactive capabilities, we further allow entity agents to take an action automatically when appro- priate. the guiding principle is that when a decision is of high confidence and low significance, the action can be car- ried out without waiting for an explicit instruction from the human operator. currently, there are three decision modes defined in the system. in the auto mode, an entity agent automatically chooses the best option available for the current situation. the recommend mode is often used when the selected option is of high consequence, wherein the agent raises a dialog prompt with a ranked list of the available options for the user. the don’t know mode is used when the system does not recognize the current situation and/or the system cannot identify an appropriate action. the mode selection rule module evaluates the decision mode for the current situation according to the ranked list of choices. if the agent cannot identify an action choice that can attain the assigned goal, the decision mode is set to don’t know. if the top choice is of high consequence, the recommend mode is used to prompt the user for a final decision. otherwise, the decision mode is set to auto, under which the agent will carry out the top action choice auto- matically without the need for user intervention. the following is a sample rule used for mode selection. this rule says that the decision mode is set to auto if the situation type is goalstatus and the agent has found that it is impossible to attain the goal with the given resource. consequently, the agent will report to the headquarter automatically that the entity is unable to attain its assigned goal. rule name = goalstatusreporttohq if decisionassist is true situationtype==goalstatus goalattainment is false then decisionmode = auto different entities can be associated with different rules. y.-h. feng et al. / expert systems with applications ( ) – author's personal copy . case study: mission on mars our implementation of cads is based on gecco (gecco, ), which is capable of supporting various scenarios, including war strategy games, fire fighting and rescue missions. a simple exploration and movement mis- sion is used in our study to demonstrate the various capa- bilities described. given a fictional mission on mars, a space vehicle explorer er is tasked to explore the unknown areas on mars. with a sensor network deployed on the explored regions, the areas are continually monitored and activities are reported to the situation model. consider a scenario in which the explorer er is required to return to its space station before running out of fuel. along the way, the explorer may encounter aliens which will threaten his survival and thus the attainment of the goal. another threat that may affect goal attainment is the changes in the terrain situation. for example, an onset of bad weather may prevent it from reaching the destination on time. in our simulated environment, there is another entity called headquarter that oversees the explorer’s mission and makes critical decisions only when necessary. during the simulation, two types of events are injected into the virtual environment: terrain events and entity events. a terrain event changes the terrain model, for instance, by blocking a particular node. an entity event leads to changes in a selected entity, either in terms of its attribute values or activities. as the situation changes upon each event injection, the situation model is constantly updated accordingly. in each cycle, the agent re-assesses the situation, gains a new understanding and possibly pro- jects future changes. as shown in the cads graphical interface (fig. ), all entities are displayed as image icons overlayed on the ter- rain map. each entity is represented and assisted by an entity agent so that the sensed data in the situation aware- ness model are processed and filtered and only relevant information are presented to the user. for each entity, its goals, missions and plans are displayed accordingly on the main panel, together with other important attributes, such as the fuel level and the health index. each entity agent communicates with its user through a message panel (in the bottom right corner), in which incoming events are displayed according to their significance, and recommenda- tions are provided to the user as the situation arises. the message panel and the graphical terrain map together enhance the user awareness of the situation in a context- aware manner. the display on the interface automatically switches into the corresponding perspective when a differ- ent entity is selected. besides classifying and displaying events in the appro- priate message levels, the entity agent pro-actively assesses the situation for available action options and ranks them accordingly. if an action is required of which the entity agent has no authority to perform, the agent raises a choice dialog window to the user, with a ranked list of the actions and its recommendations. as shown in fig. , an alien is spotted along the planned route of er . the agent detects the threat, assesses the various options, and recom- fig. . the context-aware graphical interface of cads. y.-h. feng et al. / expert systems with applications ( ) – author's personal copy mends reroute as the best solution. as reroute is a major decision, the action is posted as an recommendation to the user together with a detailed description of the analysis and reasoning process. . performance evaluation in this section, we present an experimental study in which a human operator is asked to gauge the accuracy of the entity agent acting for explorer er in terms of event classification and action recommendation. the human operator is to determine, given a specific situation, if the system classifies an incoming event or recommends an action in an appropriate manner. in addition, we esti- mate the effectiveness of the er agent in terms of the reduction in the cognitive load of its commander. during the course of simulations, events can be gener- ated based on either implicit or explicit triggers. implicit triggers are generated from the natural evolution of the simulation (e.g., the programmed responses of an entity towards the others), whereas explicit triggers are the results of external event injection. we use the following two key parameters in generating the test scenarios: (i) initial posi- tion: the initial position of er determines its route selection and therefore influences the subsequent evolution of the simulation; and (ii) terrain change: any variation in the terrain data, for example the blocking of a road and the clearing of a blocked road, may induce the entity into mak- ing the necessary adjustment to its current plan. each distinct initial position of er marks a unique scenario setting. in each scenario, multiple situations are created and decisions are made as the scenario plays out. a total of situations based on five scenarios have been collected. the goal of er remains unchanged through- out all the scenarios. . . event classification a primary role of entity agents is to classify incoming events into three levels of significance, namely information, warning and alert. we define two performance measures for this function, namely the accuracy of event classifica- tion and the reduction in the cognitive load for the human operators. based on the test situations, a human operator is asked to gauge the accuracy of the system in terms of event clas- sification. when the message level assigned by cads to an event is different from that of the operator, the specific event classification is deemed as inappropriate. as shown in table , the entity agent achieves a high accuracy of . % in classifying events. this high level of performance is achievable as the task of classifying events is relatively straightforward and our rules are sufficiently rich to cover most types of events. context-based classification of events contributes to the reduction of cognitive load on the human operators. instead of having to attend to each and every incoming events, a user now only needs to pay attention to events classified as warning and alert. to evaluate the system’s performance in quantitative terms, we define the cognitive load reduction (clr) index for an event type c formally fig. . an illustration of context-based decision support. table the prediction accuracy in event classification number of predictions number of correct prediction accuracy (%) information . warning . alert . total . y.-h. feng et al. / expert systems with applications ( ) – author's personal copy as rc ¼ �ðn c=nÞ, where nc is the number of event c requiring attention and n is the total number of incoming events. as shown in table , the entity agent has been effec- tive in reducing the cognitive load of the er com- mander by . % for warning and . % for alert. . . action recommendation in this set of the experiments, we compare the recom- mendations made by the entity agent against the preferred action choices of the human operator. as observed in table , among the actions chosen, the entity agent has a prediction accuracy of . %, . % and . % for increase speed, resume and reroute, respectively. fig. shows an instance of the incorrect action recommenda- tions. the resume action was suggested to the human oper- ator when in fact the reroute option would be the more appropriate one. as the scenario evolved, the explorer er was observed to sustain attack by the alien. we note that not all the action choices are used during the experiments. this indicates that the specified rules have not been able to cover the space of the available actions adequately, especially on the aspect of controlling speed. potentially, more rules can be added to cover this defi- ciency. however, the level of system complexity increases as the decision space gets more extensive with the increased numbers of rules and situation parameters. in particular, as different initial conditions may give rise to a variety of sit- uations as the scenario evolves, the task of addressing every single possible situation is a great challenge. . related work intelligent agents have been used for situation awareness modelling in a number of prior work. urlings, tweedale, sioutis, and ichalkaranje ( ) employ bdi (beliefs– desires–intentions) agents to enhance situation awareness of human–machine team in a military environment. based on the jack agent development environment and the unreal tournament (ut) game engine, they present a multi-agent framework. three types of agents are described, including a commander agent for supporting the human commander, a communication agent for com- munication between the commander agent and the human commander, and troop agents for exploring the environ- ment and for executing the orders. although their com- mander agent is similar to our entity agents, urling et al. are focusing on situation awareness in single agent and they table the clr indices for warning and alert events total number of events number of warnings number of alerts clri . % . % table the prediction accuracy in action recommendation action choice number of predictions number of matches prediction accuracy (%) increase speed . decrease speed na resume . reroute . wait na total . fig. . an incorrect prediction by cads. y.-h. feng et al. / expert systems with applications ( ) – author's personal copy do not employ the concept of shared situation awareness. in addition, although a high level description of the com- mander agent is provided, there is no concrete account on the specific decision support functions implemented and their evaluation. so and sonenberg ( ) incorporate the notion of situ- ation awareness into a computational model for proactive agent behavior. their model is based on a rule-based knowledge representation and a forwarding reasoning mechanism. while they develop a meta-control strategy for directing an agent’s attention in sensing and delibera- tion, their focus again is on single agent and the approach is to build a situation awareness model for each agent. more recently, thangarajah, padgham, and sardina ( ) also adopt the notion of ‘situation’ for decision making in intelligent agents. instead of using the endsley’s three-level hierarchy, they treat situations as a conceptual entity which arguably allows a richer semantics specifica- tion. for modelling the agents’ reasoning process, than- garajah et al. extend a bdi language called can for rule-based specification of the agents’ behavior. although they see the need of organizing situation objects according to their relevance to individual agents for efficiency reason, their system again does not have the notion of shared situa- tion awareness and context-aware decision support. . conclusions we have shown how context-awareness can be exploited in situation assessment and how context-aware decision support can be used to reduce the cognitive load of human operators in the command and control domain. compared with traditional situation awareness models, our system provides customized views and ser- vices out of a shared situation awareness so as to sup- port focused and effective situation awareness based on context. more importantly, we illustrate that a context- based inference engine can be used to support a myriad of proactive decision support functions that facilitate commanders to operate in complex and time critical operations. whereas endsley has also advocated a goal-oriented approach to designing situation models, we provide a computational model as well as its imple- mentation for exploiting goal-based contextual informa- tion to achieve user-specific situation awareness. while the current decision support engine based on pro- duction rules has served the purpose of knowledge acquisi- tion and the proof of concepts, the task of defining the necessary heuristics based on a bounded definition of the command and control application and responding to each and every new development can be tedious. instead of using rule-based specification, machine learning capabili- ties can be incorporated to aid in expanding the scope of the decision-making module. in fact, the task of incorpo- rating learning for situation awareness and context-based decision support has been identified as our next research target. ultimately, we are moving towards a cognitive deci- sion system that will support both direct rule-based knowl- edge specification as well as learning based knowledge acquisition based on human-agent interaction. although our experimentation thus far is based on a rel- atively simple scenario of explorer movement and alien avoidance, the framework theoretically can be applied to the general domain of command and control, including battlefield modelling and tactical warfare planning. these will form part of our future work. acknowledgements the reported work is supported in part by a research grant from the intelligent systems centre. the authors thank jun jin and chun-yip lam for contributing to the development of the mission on mars simulator. references dey, a. k. ( ). understanding and using context. personal and ubiquitos computing, , – . endsley, m. r. ( ). design and evaluation for situation awareness enhancement. in proceedings of the human factors society nd annual meeting, santa monica, ca (vol. , pp. – ). endsley, m. r. ( ). toward a theory of situational awareness in dynamic systems. human factors, , – . endsley, m. r. ( ). designing for situation awareness in complex systems. in proceedings of the second international workshop on symbiosis of humans, artifacts and environment, kyoto, japan. gecco ( ). game environment for command and control operations. available from:. harter, a., hopper, a., steggles, p., ward, a., & webster, p. ( ). the anatomy of a context-aware application. in proceedings of the th annual acm/ieee international conference on mobile computing and networking, seattle, wa (pp. – ). moran, t. p., & dourish, p. ( ). introduction to this special issue on context-aware computing. human computer interaction, , – . rupp, n. a. ( ). an introduction to the drools project. available from: . schilit, b., & theimer, m. ( ). disseminating active map information to mobile hosts. ieee network, , – . so, r., & sonenberg, l. ( ). situation awareness in intelligent agents: foundations for a theory of proactive agent behavior. in proceedings of ieee/wic/acm international conference on intelligent agent technology (pp. – ). thangarajah, j., padgham, l., & sardina, s. ( ). modelling situations in intelligent agents. in proceedings of fifth international joint confer- ence on autonomous agents and multiagent systems, hakodate, japan (pp. – ). urlings, p., tweedale, j., sioutis, c., & ichalkaranje, n. ( ). intelligent agents and situation awareness. in proceedings of the th international conference on knowledge-based intelligent information and engineering systems, lncs (pp. – ). y.-h. feng et al. / expert systems with applications ( ) – research article mathematical modeling of the expert system predicting the severity of acute pancreatitis maria a. ivanchuk, vitalij v. maksimyuk, and igor v. malyk department of biological physics and medical informatics, bukovinian state medical university, kobyljanska street , chernivtsi , ukraine department of surgery, bukovinian state medical university, golovna street , chernivtsi , ukraine department of the system analysis and insurance and financial mathematics, chernivtsi national university of yuriy fedkovich, unversitetska street , chernivtsi , ukraine correspondence should be addressed to maria a. ivanchuk; mgracia@ukr.net received december ; accepted may ; published june academic editor: daniel kendoff copyright © maria a. ivanchuk et al. this is an open access article distributed under the creative commons attribution license, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. the method of building the hyperplane which separates the convex hulls in the euclidean space 𝑅𝑛 is proposed. the algorithm of prediction of the presence of severity in patients based on this method is developed and applied in practice to predict the presence of severity in patients with acute pancreatitis. . introduction during the last decades, pronounced tendency to the relentless increase in morbidity in acute pancreatitis is observed. thus, the depth of pathomorphological pancre- atic parenchyma lesions can vary from the development of edematous pancreatitis up to pancreatic necrosis. however, accurate predicting of the probable nature of the lesion of the pancreas in the early stages of acute pancreatitis is one of the most difficult problems of modern pancreatology. diagnostic and the predictive probability of existing laboratory and instrumental diagnostic markers and rating scales does not exceed – % [ – ]. such situation is a major difficulty in selecting the adequate treatment strategy in the initial stages of acute pancreatitis. thus the search for new methods of accurate predicting of acute pancreatitis’ severity becomes an urgent problem. development of mathematical approaches for prediction in medicine was developed by fisher, the father of the linear discriminant analysis [ ]. currently, there are many approaches to solving this problem: cluster analysis, the construction of predictive tables, image recognition, and lin- ear programming. fundamentals of building the prognostic tables and wald serial analysis are described in [ ]. cluster analysis is commonly used for solving the tasks of medical prediction. in the paper [ ], the procedure of cluster analysis with a study of the indices of the daily variability of cardiac rhythm in patients with the ischemic disease of heart is examined. in [ ] using national data from the scientific registry of transplant recipients authors compare transplant and wait-list hospitalization rates. they suggest two marginal methods to analyze such clustered recurrent event data; the first model postulates a common baseline event rate, while the second features cluster-specific baseline rates. results from the proposed models to those based on a frailty model were compared with the various methods compared and contrasted. three major considerations in designing a cluster analysis are described in [ ]. the first relates to selection of the individuals. the second consideration is selection of variables for measurement and the third consideration is how many variables to choose to enter into a cluster analysis. to classify clinical phenotypes of anti-neutrophil cytoplasmic antibody-associated vasculitis, cluster analysis was used in [ ]. researches on the general theory of diagnosis, classifi- cation, and application of optimization methods for pattern recognition, solving applied problems in medicine and biol- ogy, are conducted by mangasarian et al. for many years [ ]. hindawi publishing corporation journal of computational medicine volume , article id , pages http://dx.doi.org/ . / / journal of computational medicine but universal method for solving problems of recogni- tion, identification, and diagnosis does not exist. therefore, development of methods for predicting in medicine still remains relevant. one among the many challenges of recog- nition is the task of constructing hyperplanes which separate two convex sets. many manuscripts [ – ] are devoted to the solution of this problem. we propose a methodology for constructing convex hulls and their separation, which can be used for modeling expert medical prognostic systems (e.g., to separate groups of patients with different degrees of severity of the disease for prediction of severity in patients). . methods . . separation of the convex hulls. let us have two sets of points 𝐴 = {𝑎 𝑖 = (𝑎 𝑖 ,𝑎 𝑖 , . . . ,𝑎 𝑛 𝑖 ), 𝑖 = ,𝑚 𝐴 } and 𝐵 = {𝑏 𝑖 = (𝑏 𝑖 ,𝑏 𝑖 , . . . ,𝑏 𝑛 𝑖 ), 𝑖 = ,𝑚 𝐵 } in euclidean space 𝑅𝑛. let 𝑚 be number of points in the set. we must find the separate hyperplane: 𝐿 𝑝 = {𝑥 ∈𝑅 𝑛 : ⟨𝑝,𝑥⟩ = 𝛾}, 𝑝 ̸= , ( ) where ⟨𝑝,𝑥⟩ is the scalar product of the vectors𝑝and𝑥 such that sets 𝐴 and 𝐵 can be placed in the different half-spaces: 𝐿 + 𝑝 = {𝑥 ∈𝑅 𝑛 : ⟨𝑝,𝑥⟩ > 𝛾}, 𝐿 − 𝑝 = {𝑥 ∈𝑅 𝑛 : ⟨𝑝,𝑥⟩ < 𝛾}. ( ) to build the convex hull conv 𝐴 for the set 𝐴, for each of 𝐶 𝑛 𝑚 𝐴 points’ combinations from the set𝐴, if it is possible, build the hyperplane 𝐻 𝑝 = {𝑥 ∈𝑅 𝑛 : ⟨𝑝,𝑥⟩ = 𝛽}, 𝑝 ̸= . ( ) coordinates of the vector 𝑝 = (𝑝 , . . . ,𝑝𝑛) are found as minors(𝑛− )order for elements of the first row of the matrix: ( 𝑥 −𝑎 𝑥 −𝑎 ⋅ ⋅ ⋅ 𝑥 𝑛 −𝑎 𝑛 𝑎 −𝑎 𝑎 −𝑎 ⋅ ⋅ ⋅ 𝑎 𝑛 −𝑎 𝑛 ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ ⋅ 𝑎 𝑛 −𝑎 𝑎 𝑛 −𝑎 ⋅ ⋅ ⋅ 𝑎 𝑛 𝑛 −𝑎 𝑛 ), ( ) where 𝑥 ∈ 𝑅𝑛, 𝑎 𝑖 ∈ 𝐴, 𝑖 = ,𝑛. coefficient 𝛽 is determined from the following equation: 𝛽=−(𝑎 𝑝 +𝑎 𝑝 + ⋅ ⋅ ⋅ +𝑎 𝑛 𝑝 𝑛 ). ( ) if all points of the set 𝐴 are in the one of half-spaces of hyperplane 𝐻 𝑃 , then polygon 𝑎 𝑎 ⋅ ⋅ ⋅𝑎 𝑛 is one of the convex hull’s hyperfaces. the complex of all hyperfaces is the convex hull conv 𝐴 . point 𝑏 𝑖 ∈ 𝐵 is called outlier if point 𝑏 𝑖 is internal for the conv 𝐴 . point 𝑏 𝑖 is outlier if there is at least one hyperface 𝑎 𝑎 ⋅ ⋅ ⋅𝑎 𝑛 ∈ conv 𝐴 that 󵄨󵄨󵄨󵄨󵄨󵄨󵄨 󳨀→ 𝑐𝑓 󵄨󵄨󵄨󵄨󵄨󵄨󵄨 = 󵄨󵄨󵄨󵄨󵄨󵄨󵄨 󳨀→ 𝑐𝑏 𝑖 󵄨󵄨󵄨󵄨󵄨󵄨󵄨 + 󵄨󵄨󵄨󵄨󵄨󵄨󵄨 󳨀󳨀→ 𝑏 𝑖 𝑓 󵄨󵄨󵄨󵄨󵄨󵄨󵄨 , ( ) where point 𝑐 ∈ int conv 𝐴 , point 𝑓 is the intersection point of the hyperplane 𝐻 𝑝 (𝑎 𝑎 ⋅ ⋅ ⋅𝑎 𝑛 ∈𝐻 𝑝 ), and line 𝑥 −𝑐 𝑏 −𝑐 = 𝑥 −𝑐 𝑏 −𝑐 = ⋅ ⋅ ⋅ = 𝑥 𝑛 −𝑐 𝑛 𝑏 𝑛 −𝑐 𝑛 . ( ) to find the point 𝑓 let us write ( ) in parametric form: 𝑥 = 𝑐 +(𝑏 −𝑐 )𝑡 𝑥 = 𝑐 +(𝑏 −𝑐 )𝑡 ⋅ ⋅ ⋅ 𝑥 𝑛 = 𝑐 𝑛 +(𝑏 𝑛 −𝑐 𝑛 )𝑡. ( ) put ( ) in the hyperplane equation ( ) and find parameter 𝑡: 𝑡 ∗ = 𝑝 𝑐 +𝑝 𝑐 + ⋅ ⋅ ⋅ +𝑝 𝑛 𝑐 𝑛 +𝛽 𝑝 (𝑐 −𝑏 )+𝑝 (𝑐 −𝑏 )+ ⋅ ⋅ ⋅ +𝑝 𝑛 (𝑐 𝑛 −𝑏 𝑛 ) . ( ) to find coordinates of the point 𝑓 let us put ( ) in ( ): 𝑓 = 𝑐 +(𝑏 −𝑐 )𝑡 ∗ 𝑓 = 𝑐 +(𝑏 −𝑐 )𝑡 ∗ ... 𝑓 𝑛 = 𝑐 𝑛 +(𝑏 𝑛 −𝑐 𝑛 )𝑡 ∗ . ( ) after finding all outliers from the sets 𝐴 and 𝐵 eject outliers from the set, with less number of outliers. build the new convex hulls and find the outliers. if there are outliers in the new convex hulls, eject them. if there are not any outliers, the convex hulls do not intersect. according to consequence of hahn-banach theorem there is a nonzero linear functional 𝐿 𝑝 that separates conv 𝐴 and conv 𝐵 [ ]. find the separating functional 𝐿 𝑝 as hyperplane parallel to one of convex hulls’ hyperfaces. choose hyperface so that convex hulls conv 𝐴 and conv 𝐵 are in different half-spaces formed by hyperplane parallel to this hyperface. find points 𝑎min ∈ conv𝐴 and 𝑏min ∈ conv𝐵 so that | 󳨀󳨀󳨀󳨀󳨀󳨀→ 𝑎min𝑏min| = min(| 󳨀󳨀→ 𝑎 𝑖 𝑏 𝑗 | : 𝑎 𝑖 ∈ conv 𝐴 , 𝑏 𝑗 ∈ conv 𝐵 , 𝑖 = ,𝑚 𝐴 , 𝑗 = ,𝑚 𝐵 ). let 𝑑min ∈ 󳨀󳨀󳨀󳨀󳨀󳨀→ 𝑎min𝑏min. for each hyperface {𝐻𝐴min : 𝐻 𝐴min ⊂ conv 𝐴 ;𝑎min ∈ 𝐻𝐴min} and {𝐻𝐵min : 𝐻𝐵min ⊂ conv 𝐵 ;𝑏min ∈ 𝐻𝐵min} build the parallel hyperplane {𝐿𝑝 : 𝑑min ∈ 𝐿𝑝;𝐿𝑝‖𝐻𝐴minor 𝐿𝑝‖𝐻𝐵min}. if 𝑎𝑖 ∈ 𝐿 + 𝑝 , for all 𝑎 𝑖 ∈ 𝐴, 𝑖 = ,𝑚 𝐴 , and 𝑏 𝑗 ∈ 𝐿 − 𝑝 , for all 𝑏 𝑗 ∈ 𝐵, 𝑗 = ,𝑚 𝐵 , then 𝐿 𝑝 is separating hyperplane. . . modeling the expert system of predicting the presence of severity in patients. let us have two groups of patients: 𝐴, patients with severity, and 𝐵, patients without severity. there are 𝑛 parameters (factors which affect the severity) known for each patient. journal of computational medicine during modelling we used the terms sensitivity (se) and specificity (sp): se = 𝑎 𝑎+𝑐 , sp = 𝑑 𝑏+𝑑 , ( ) where𝑎 is the true positives,𝑏 is the false positives (overdiag- nosis errors), 𝑐 is the false negatives (underdiagnosis errors), and 𝑑 is the true negatives. the sensitivity of a clinical test refers to the ability of the test to correctly identify those patients with the disease. the specificity of a clinical test refers to the ability of the test to correctly identify those patients without the disease [ ]. we created an algorithm of modelling the expert system in a way that uses the least amount of features for the best result. information of the parameters was found using kulback’s information measure [ ]. we built convex hulls for the most informative factor. if convex hulls intersect, we found outliers—the points from the set 𝐴 that are internal to conv 𝐵 and the points from the set𝐵 that are internal to conv 𝐴 . the set𝐴outliers are underdiagnosis errors. the set𝐵outliers are overdiagnosis errors. we built the prognostic system to find the patients with severity, so we rejected the outliers from the set𝐵. let the set𝑂 𝐵 = {𝑜 𝑖 : 𝑜 𝑖 ∈𝐵∩ int conv 𝐴 , 𝑖 = ,𝑚 𝑂 𝐵 } be the set of outliers from 𝐵. after rejecting, we get a new set 𝐵 󸀠 =𝐵/𝑂 𝐵 . if you build the expert system for differential diagnosis, you reject outliers out of the set where there are less of them. if the percentage of rejected points is more than the significance level 𝑚 𝑂 𝐵 𝑚 𝐵 >𝛼, ( ) the next (the most informative) factor was added. the space dimension is increased by . in the new space convex hulls were built and the outliers were rejected. the space dimension was increased until preassigned significance level. if all available diagnostic information was used, but preassigned significance level was not reached, then decision of not suffi- cient information was taken. when preassigned significance level was reached, we found the separating hyperplanes. the algorithm for modelling the prognostic system is represented on the figure . the results were checked in the control group and the hyperplane with maximal sensitivity was chosen. the complexity of this algorithm is 𝑂(𝑚𝑛+ ) [ ] if the convex hulls are built by search of all combinations of points. the complexity of this algorithm is 𝑂(𝑚 ) if the convex hulls are built by jarvis march or “gift wrapping” algorithm [ ]. . results . . the expert system of predicting the presence of severity in patients with acute pancreatitis. the research involved persons with severe and patients with nonsevere acute pancreatitis. among them, there were ( . %) men and ( . %) women. the mean age was . years (± . ) in males and . (± . ) in females. the most common etiology was alcohol consumption ( . %), followed by gallstones ( . %). in . % no identifiable cause was found. + − find the outliers o + − decision of not sufficient begin end reject the outliers b󳰀 = b/o find lp n := mo mb > 𝛼 n := n + n < n information choose lmaxp : se(lmaxp ) build conva, convb find convb󳰀 = max {se(lp)} figure : algorithm for modelling the prognostic system. the diagnostic criteria for acute pancreatitis were those defined by the ap guidelines, as the presence of at least two of the following features: ( ) characteristic abdominal pain, ( ) elevation over times the upper normal limit of serum amylase/lipase, and ( ) characteristic features on com- puter tomography (ct) scan [ ]. severe acute pancreatitis was diagnosed according strictly to atlanta criteria: early prognostic scores, apache ii ≥ , ranson ≥ ; organ failure, systolic pressure < mmhg, creatinine > . mg/l after rehydration, pao ≤ mmhg; local complications (on ct scan), necrosis, abscess, and pseudocyst [ ]. patients were divided into two samples—training ( patients with severity and without them) and control ( patients with severity and without). the level of significance was 𝛼 = , . the algorithm presented above was used for patients with training set. for 𝑛 = , the percentage of outliers was . %. for 𝑛 = , the percentage of outliers was %. for 𝑛 = , the percentage of outliers was . %. for 𝑛 = , the percentage of outliers was %. we got hyperplanes which separate the convex hulls of the training samples. two of them had higher sensitivity and specificity (we got only ( %) of underdiagnosis errors journal of computational medicine and there were no overdiagnosis errors for the control sample with these hyperplanes): − . 𝑥 − . 𝑥 + . 𝑥 + 𝑥 + . = , − . 𝑥 − . 𝑥 + . 𝑥 + . 𝑥 + . = , ( ) where 𝑥 is time before hospitalization, 𝑥 is blood lipase, 𝑥 is amylase urine, and 𝑥 is bmi. so, we built the expert system with sensitivity se = %. statistical errors are seen only in patient in the con- trol group, who were diagnosed with interstitial edematous pancreatitis development on the background severe diabetes mellitus. according to expert system the acute pancreatitis without severity was predicted. this error, in our view, is associated with late ambulation of the patient for medical care as a result of atypical course of acute pancreatitis, increased blood and urine amylase, and increased bmi, which is the characteristic signs of diabetes mellitus. that is, in this case, some of the most important prognostic parameters of acute pancreatitis have been characterized by different diseases in particular of diabetes mellitus, which caused the error. . conclusions the method of separation of convex hulls in euclidean space by constructing a separating hyperplane parallel to one of the convex hulls hyperfaces is proposed. on the basis of this method the algorithm for modelling the prognostic system is stated. the proposed algorithm is applied in practice to predict the presence of severity in patients with acute pancreatitis and gives % correct results for the control sample, while diagnostic and the predictive probability of existing laboratory and instrumental diagnostic markers and rating scales does not exceed – %. clinical application of the developed mathematical model predicting the severity of acute pancreatitis promotes proper choice of treatment tactics and allows improving final results of these patients’ treatment. conflict of interests the authors declare that there is no conflict of interests regarding the publication of this paper. references [ ] e. j. balthazar, “acute pancreatitis: assessment of severity with clinical and ct evaluation,” radiology, vol. , no. , pp. – , . [ ] g. sathyanarayan, p. k. garg, h. k. prasad, and r. k. tandon, “elevated level of interleukin- predicts organ failure and severe disease in patients with acute pancreatitis,” journal of gastroenterology and hepatology, vol. , no. , pp. – , . [ ] a. k. khanna, s. meher, s. prakash et al., “comparison of ran- son, glasgow, moss, sirs, bisap, apache-ii, ctsi scores, il- , crp, and procalcitonin in predicting severity, organ failure, pancreatic necrosis, and mortality in acute pancreatitis,” hpb surgery, vol. , article id , pages, . [ ] r. a. fisher, contributions to mathematical statistics, john wiley & sons, . [ ] e. v. gubler and a. a. genkin, primenenie neparametricheskih kriteriev statistiki v medico-biologicheskih issledovanijah (using the nonparametric statistic criteria in medical and biological researches), leningrad, russia, . [ ] a. i. borodulin, a. v. sviridov, and o. v. sudakov, “using the cluster analysis in the in the diagnostic of ischemic heart disease,” prikladnie informacionnie aspekti medicini, no. , pp. – , . [ ] d. e. schaubel and j. cai, “analysis of clustered recurrent event data with application to hospitalization rates among renal failure patients,” biostatistics, vol. , no. , pp. – , . [ ] m. weatherall, p. shirtcliffe, j. travers, and r. beasley, “use of cluster analysis to define copd phenotypes,” european respiratory journal, vol. , no. , pp. – , . [ ] a. mahr, s. katsahian, h. varet et al., “revisiting the classi- fication of clinical phenotypes of anti-neutrophil cytoplasmic antibody-associated vasculitis: a cluster analysis,” annals of the rheumatic diseases, vol. , no. , pp. – , . [ ] o. l. mangasarian, w. n. street, and w. h. wolberg, “breast carter diagnosis and prognosis via linear programrnig,” opera- tion research, vol. , no. , . [ ] a. i. golikov, y. g. evtushenko, and s. ketabchi, “on families of hyperplanes that separate polyhedra,” computational math- ematics and mathematical physics, vol. , no. , pp. – , . [ ] a. r. alimov and v. y. protasov, “separation of convex sets by extreme hyperplanes,” fundamental and applied mathematics, vol. , no. , pp. – , . [ ] i. i. eremin, “fejér methods for the strong separability of convex polyhedral sets,” russian mathematics, no. , pp. – , . [ ] a. v. ershova, “algorithm of the disjoint convex polyhedral separating using fejer mappings,” sistemi upravlenia i informa- cionnie tehnologii, vol. , no. , pp. – , . [ ] c. böhm and h.-p. kriegel, “determining the convex hull in large multidimensional databases,” in data warehousing and knowledge discovery, vol. of lecture notes in computer science, pp. – , springer, . [ ] g. fung, r. rosales, and b. krishnapuram, “learning rankings via convex hull separation,” in proceedings of the annual conference on neural information processing systems (nips ’ ), pp. – , december . [ ] a. n. kolmogorov and s. v. fomin, elements of the theory of functions and functional analysis, courier dover, . [ ] d. g. altman, practical statistics for medical research, chapman & hall/crc, london, uk, . [ ] s. a. abramov, lectures at algorithmic complexity, mtsnmo, moscow, russia, . [ ] f. preparata and m. shamos, computational geometry—an introduction, springer, . [ ] p. a. banks and m. l. freeman, “practice guidelines in acute pancreatitis,” the american journal of gastroenterology, vol. , no. , pp. – , . [ ] e. l. bradley iii and c. f. frey, “a clinically based classification system for acute pancreatitis: summary of the international symposium on acute pancreatitis, atlanta, ga, september through , ,” archives of surgery, vol. , no. , pp. – , . submit your manuscripts at http://www.hindawi.com stem cells international hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume mediators inflammation of hindawi publishing corporation http://www.hindawi.com volume behavioural neurology endocrinology international journal of hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume disease markers hindawi publishing corporation http://www.hindawi.com volume biomed research international oncology journal of hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume oxidative medicine and cellular longevity hindawi publishing corporation http://www.hindawi.com volume ppar research the scientific world journal hindawi publishing corporation http://www.hindawi.com volume immunology research hindawi publishing corporation http://www.hindawi.com volume journal of obesity journal of hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume computational and mathematical methods in medicine ophthalmology journal of hindawi publishing corporation http://www.hindawi.com volume diabetes research journal of hindawi publishing corporation http://www.hindawi.com volume hindawi publishing corporation http://www.hindawi.com volume research and treatment aids hindawi publishing corporation http://www.hindawi.com volume gastroenterology research and practice hindawi publishing corporation http://www.hindawi.com volume parkinson’s disease evidence-based complementary and alternative medicine volume hindawi publishing corporation http://www.hindawi.com an expert system for modelling wave - height time - series an expert system for modelling wave - height time - series rodolfo piscopia freelance rome, italy abstract— this paper describes an expert system designed for the analysis of an incomplete, non-stationary and non- gaussian, long-term, time series of wave significant heights by means of specific linear parametric model. using this system makes it possible to complete missing-value gaps, forecast wave- height short-term evolution or simulate arbitrarily long sequences of wave data preserving the key statistical properties of the observed series, including autocorrelation, persistence over threshold, non-gaussian distribution and seasonality. the implemented improvements bear on specific key tasks of arma setup procedure, i.e. preliminary analysis, parameter estimation and optimal model-configuration identification. specifically, a seasonal trend decomposition based on loess robust method is applied to compute more stable and detailed seasonal trend, allowing assuming more confidently its deterministic nature. moreover, aiming at accurately estimating the model parameters, a proficient method is taken in, which is based on the robust whittle’s approximation of the maximum log-likelihood function as well as on the direct-search, non- linear, multi-parameter, constrained, optimization technique called complex modified. finally, an automatic expert system is developed, able to identify, almost correctly, arma orders by selecting the model with the smallest residuals variance and parameter numbers. confident applicability of the suggested procedure is tested by means of both monte carlo simulations and comparisons of generated series with observed one, this latter measured offshore alghero – italy. analysis of results clearly indicate that the accuracy in identifying the correct arma model is improved; furthermore, it is shown that the simulated time series exhibit all the primary statistical properties of the observed data, including winter and summer seasonal patterns as well as sea states sequencing, persistence and severeness. keywords — wave climate; arma model; wave forecast; storm duration; sea state persistence; sea severeness i. introduction for marine human activities and engineering applications, the understanding of sea-state sequences is important as well as the knowledge of extreme wave parameters, e.g. to evaluate a maritime traffic line efficiency, to guess a port/terminal operativeness or to assess risks of engineering processes. actually, marine intervention and installation works involve long-lasting and complex operations. in these cases, the analysis of effects related to meteorological changes during specific operations is utmost relevant to disclose any possible critical situations and their related costs-growth. to these aims, linear models can be very helpful, being able to provide large database of information statistically equivalent to the observed one. additionally, recorded time series are usually incomplete due to several reasons, e.g. to instrument failures, accidental data loss or spikes rejection. considering that the data incompleteness can seriously bias statistical inferences, makes obvious the relevance of a procedure able to recover missing values by ensuring same statistical sample properties. autoregressive, moving-average models (arma) are a specific class of the linear parametric family that, in few words, replicate time processes by combining some their outcomes with a white noise. in ocean engineering applications [ ] and [ ] have used arma to simulate individual waves in short-term elevation record, supposed to be stationary in time. reference [ ] have used arma to model the non-stationary, long-term, time- series of significant wave-height, whereas [ ] proposed a new methodology for the analysis, missing-value completion and simulation of an incomplete, non-stationary, time-series of wave data. further researches were pointed at verifying data transferability between two wave-measuring stations [ ]. generally, two main problems have to be solved in order to apply arma models to long-term series of wave parameters. one is the presence in the series of missing-value gaps, which can sometimes be relatively long, and the other is the series non-stationarity and non-normality. accordingly, gap filling as well as data transformation procedures are required. furthermore, the common and challenging task of the model identification, i.e. selecting the most suitable arma order, has to be tackled. here, the work of [ ] is extended by improving the following tasks: the seasonal component assessment, the model parameter estimation and the choice of the optimal arma configuration. namely, a technique called stl robust (seasonal trend decomposition based on loess) able to compute more accurate seasonal components is adopted [ ]. furthermore, the more robust whittle’s approximation of the maximum log-likelihood function is used to estimate arma coefficients, the set of which is found out by a proficient, nonlinear, constrained, multi-parameter, optimization technique. finally, an expert system has been developed allowing automating the struggle step of model identification that, for mixed arma process, is quite tricky and someway affected by subjective interpretation. different monte carlo simulations have been carried out with a double purpose: the validation of the parameter estimation procedure and the verification of the automatic expert system proficiency. the obtained results have been gratifying, making possible to confidently say that the proposed enhancements are very efficient. international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- in the following, each step of the adopted arma modelling procedure is accurately described and the results of monte carlo simulations are illustrated. afterwards, the application to real wave data is fully described and the comparison between two different techniques to normalize- denormalize the series is plainly outlined. finally, the conclusions are drawn out. ii. linear parametric models starting from the box and jenkins definition [ ], the family of linear models has been developed with the conception of several subtypes that roughly follow a common setup procedure. here only the arma model is considered. regarding a second-order stationary and gaussian time series zt, the autoregressive and moving average parts of an arma(p,q) model define zt respectively as the combination of p previous terms of the series plus the combination of q+ terms of a white noise (i.e. a stationary random process with zero mean and variance equal to r ). introducing the back- shift operator defined as ntt n zzbb : , an arma(p, q) can be written as     tt abzb   , being pp bbb   .... )( and q q bbb   .... )( . the standard arma setup procedure can be resumed as follows [ ]: preliminary analysis, model identification, parameter estimation, model verification and optimal model- configuration selection. in what follows each task is accurately delineated. iii. preliminary analysis with reference to the stationarity of significant wave- height series, it is typically not satisfied and, according to the common knowledge of the environmental process, a seasonal component is expected to exist. according with [ ], a non- stationary time series zt can be decomposed as ttttt xzz   ~ , where tz ~ , t , t are the deterministic functions, respectively, of long-term trend, seasonal mean and standard deviation. in what follows, the long-term trend is not considered. the seasonal mean and variance can be defined as follows by introducing the buys-ballot double index, i.e. by re- indexing the time series zt as [ ]:      y j j t mzntz , ,          y j jz k   m          y j jz k    m   where y and m are integers respectively equal to the series length in year and the annual number of observations, n=ym is the series numerosity,  is the index within the annual cycle, k is the number of existing values per each observation index  (if no missing values affects the wave series then clearly k=y). the deseasonalized series is computed according to the following expression:      jj zy with yj  and m   this approach, however, produces seasonal components possibly affected by large sample variability, especially when the time series length is not enough extended or when many missing-value gaps exist. this large variability contrasts with the assumed deterministic nature of the seasonal component and cannot therefore be accepted by both physical and stochastic points of view. following [ ], here is preferred a more robust method, derived from the stl one (seasonal-trend decomposition based on loess). this technique, used for both the mean and variance seasonal components, is split into five tasks (here, only the evaluation of the mean component is illustrated as the variance computation is straightforwardly derivable). . identify the seasonal mean series * by ( ) and reduce it to zero average. compute the new time series *  jj zx . . define the scaling factor   cxu jj  , where  is the median value of jx and c is a constant (equal to and respectively for the mean and variance seasonal components). . estimate the weighting factor j , for each jx , according to:            - j jj j uif uifu      . evaluate the weighted seasonal mean component, for each index  in the annual cycle, as:     y j j y j jjxm *     . smooth *m by means of the interpolator called loess. specifically, considering a time window centered at , with amplitude w, *m is smoothed according to:       **ˆ w wi iimm     with )(                   w i i     for completing the missing-value gaps, the following procedure has been adopted. when the gap length is very small (dealing with one or two observations), the missing values are interpolated from neighbors. otherwise, the smoothed mean seasonal component ( ) is transformed in a fourier series. the missing values are therefore replaced by [ ]:      sincos ~ baz j    being  the average value of the smoothed mean seasonal component, = /m, a and b the fourier coefficients given by:      m m a cos         m m b sin      international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- with reference to normality of the significant wave-height series, it is verified by performing a t-student test; if the tested hypothesis is rejected, a series transformation is applied. two different approaches were implemented and compared: the first involves the classical box-cox transformation [ ]; the second entails the probability level-equivalence transformation (plet) used by [ ]. using the box-cox formula, the time series is transformed as:          log )(ˆ bct bcbct bct whenz whenz z bc       differently, the plet method is based on the percentile equivalence among the standardized normal distribution () and the wave-height best-fitting distribution (pz). the standardized normal series is therefore obtained by:    tzt zpz ˆ    the inverse transformation used in the simulation task is:    tzt zpz ˆ     iv. model parameters estimation this task is here completed in two steps: the preliminary estimation and the accuracy refining. the method of moments [ ] is adopted for the former, whereas the maximum log- likelihood method along with the whittle’s function approximation is implemented for the latter. actually, for a gaussian stationary process, the approximated expression of log-likelihood function is [ ]:     zaz n dszl tw )( ,log )|( ~ ψψψ          being =( ,…,p, ,…,q, r) the parameters vector to estimate, () the inverse of the process covariance-matrix and s(, ) the parametric power spectrum. the latter is the fourier transform of the autocorrelation function and can be seen as expressing the energy level of each periodicity composing the time series. its expression, computed by parametric method, can be written as [ ]:                      p j jj q i ii r jj ii s sin cos sin cos ;    ψ   equation ( ) can be rewritten in terms of the series periodogram (p), i.e. the series power-spectrum computed by fourier method, as follows [ ]:                           d s p dszlw , ,log )|( ~ ψ ψψ   the maximization of ( ) is here carried out by a direct- search, non-linear, multi-parameter, constrained, optimization technique called complex modified [ ] [ ], which has been proved to perform very efficiently [ ]. aiming to enlighten the proficiency of the maximum log- likelihood method, two set of tests were carried out. in the first one, three different spectra were firstly defined by assigning p, q, r, j and  in ( ) and then randomly perturbed by adding a white noise drawn out from u[- . r; . r]. the resulting frequency distributions were assumed as periodograms to be fitted by the complex modified method with ( ) as target function. fig. shows the results along with those achieved using the spectral least-square target- function, given by [ ]:                     d p ps ; ψ   the fig. and the herein reported table clearly state that, even if both ( ) and ( ) fulfil the optimization by producing a nearly perfect fitting of arma spectra, ( ) is more efficient to accurately estimate each parameter value, reducing the maximum relative approximation-error form % to % (for ar( ) – ar( )) or from % to % (arma( , )). in the second test case, a monte carlo simulation was carried out by modelling synthetic series, generated from an arma( , ) with r = . and both (,  ranging from . to . , step . . for each generated series, ( ) was used to estimate the arma parameter values. the difference between the assigned values and the finally estimated one have been represented, in fig. , as relative errors in a box-plot form. the relative errors obtained by using the widespread method of moments [ ] along with those achieved by minimizing ( ) are also reported. the illustrated results, revealing a great error variance reduction (at least halved) as well as an unbiased zero averages, confidently confirm the proficient improvement achieved by the herein implemented method. v. verification of estimated models and selection of the optimal one to verify arma stationarity and invertibility, respectively, all roots of following ( ) should lie externally to the unit circle (here imsl® zporc routine is used):    p i i i x    q i i i x   if one of the tested hypothesis is rejected the model is not considered further, otherwise a portmanteau test is completed. namely, if an arma is stationary and invertible as well as properly identified with accurately estimated parameters, the model residuals, given by                q j jj p i ii azzr     have nearly null random values. international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- model = ar( ) model = ar( ) model = arma( . ) fixed mlm lsm fixed mlm lsm fixed mlm lsm r = .  r = .  r = .  r = .  r = .  r = .  r = . r = . r = . j j j j j j j i j i j i . . . - . - . - . - . . - . . - . . - . - . - . . . . - . - . - . - . - . - . . . . . . . . - . . - . . - . - . - . - . . . . . . . . . . - - - . . . . - . . - . . - . - - - . . . . . . . . . - - - . . . - . . - . . - . . - - - . . . . . . . . . - - - . . . . . . . . . - - - . . . - . - . - . - . - . - . . . . . . . . . . . . . . . . . . . . . randomly perturbed theoretical spectrum adapted spectrum theoretical spectrum . . . . . . . . . . . . . . . randomly perturbed theoretical spectrum adapted spectrum theoretical spectrum . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . lsm target function ( ) fig. . comparison between synthetics periodograms and spectral distributions obtained by the complex modified optimization technique. international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- fixed parameter values a b c d e f g h i j k l n o p q r s t u p q  # # # # . . . . . . . . . . . . . . . .  . . . . # . . . # . . . # . . . # . . . . . . . . . . - . - . - . - . - . - . m.l.m. (whittle) least square method method of moments . . . . . . . - . - . - . - . - . - . a b c d e f g h i j k l n o p q r s t u a b c d e f g h i j k l n o p q r s t u a b c d e f g h i j k l n o p q r s t u . . . . . . . - . - . - . - . - . - . m odel m odel ˆ  i m odel m odel ˆ  i m odel m odel ˆ r rir   estimates of ar coefficients estimates of ma coefficients estimates of residual variance fig. . results of the monte carlo simulation for the arma( , ) parameter estimation obtained by means of three different techniques: the methods of moments (right panels), the spectral least square method (central panels) and the whittle’s approximation of the maximum log-likelihood function (left panels). international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- to verify this hypothesis, the ljung-box portmanteau-test is used [ ]. a weighted sum (q) of residual autocorrelation coefficients is computed according to the following expression:       s k k kn nnq    where n is the residuals number, k is the time lag, k is the autocorrelation coefficient (computed by imsl® acf routine) and s is the maximum lag (here s= is adopted). q is then compared to the quantile of a  distribution, with s degrees of freedom, at the level of probability p. if q is greater than  (s), the test is rejected and at least one of the examined autocorrelation coefficients is statistically different from zero, to the fixed significance level p. if the tested arma is stationary, invertible, with random residuals, it will be considered for the final task of optimal model selection, i.e. the choice of model order (p, q). aiming at automating this specific task, many methods based on some patterns of different acf functions have been proposed [ ]. here a different approach is used. starting from both the definition and meaning of a linear model, the more efficient configuration can be defined as the one that outlines the process correlation structure by using the lowest parameter number and, at the same time, produces random residuals with the lowest variance. considering that the latter generally decreases as the former increases, makes it necessary to choose the “optimal configuration” on the basis of statistical indices. in the present work, the following is taken into consideration [ ]:                        qp σσn qp+ n-p-q σn n-p-qbic= rrr ˆ ln ˆ ln   where ˆ r σ is the variance of series-residuals. the combination of p and q that minimize ( ) is regarded as the arma “optimal configuration”. to show the consistency of the proposed procedure, two series of tests were carried out. the first one was performed analyzing the identified ranking of ten different ar, ma and mixed arma models of arbitrary orders. the obtained results are reported in table i and support, although not on a statistical basis, the developed expert-system robustness. namely, the true model order is ranked for seven times in the first two positions and it is always high-ranked. furthermore, only one attempt gave the sum p+q of the fitted model underestimated by more than one order (settled arma( , ) – selected ma( )). the second monte carlo simulation was carried out with the goal of comparing the proposed expert-system proficiency with that of the three best-performing acf pattern-selection methods; namely, the corner, the eacf and the scan methods. the efficiency of these latter methods was found out from [ ]. the arma( , )      tt abzbb . . .  with . r , was used for simulating series of terms, which were successively analyzed. table i. results of the “optimal model identification” test for ten different ar, ma and arma models of arbitrary order. fixed identified fixed ranking p q p q bic st st st rd th nd st nd th nd the occurrence of the identified combinations (p, q) obtained by the different methods are summarized in table ii. all the methods fairly spread out the identified configurations but, in this specific case, any of them significantly underestimate the model total order (p+q). the expert system performs better than the corner method. namely, the former achieves nearly equal results in selecting the true model configuration (scoring just a % less than the latter), but it shows greater sensitivity in both recognizing the minimum model orders and identifying the correct influence of the ar and ma model parts. actually, when one of the identified model order is equal or greater than the fixed one, the expert system % of times overestimates the other one whereas the corner method underestimates it % of times. moreover, the expert system % of times bias the autoregressive character of the process with the ma one whereas the corner method makes the same error for the % of times. the esacf and scan methods have instead a worse hit percentage for the correct model identification and have the same biasing character of the corner one. on these bases, it could be stated that the implemented expert system works nicely well, slightly better than the best performing pattern selection method here considered. moreover, it has to be highlighted that the expert system automatically provides in output a list of ranked models, opening to chances of trying different configurations having similar statistical index values. table ii. number of identified combinations (p, q) for the series generated from an arma ( , ). expert system corner method esacf method scan method ( , ) - - - - - - - - ( , ) - - - - - - - - ( , ) % - - - - % ( , ) - - - - - - - - ( , ) % % - - % ( , ) - - - - - - - - ( , ) - - - - - - - - ( , ) % % % % ( , ) % % % % ( , ) % % % % ( , ) % % % % ( , ) % % % % ( , ) % % % % ( , ) % % % % ( , ) % - - - - - - international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- table iii. rates of under-specification of the total arma order.   expert system % % % % % % corner method % % % % % % esacf method % % % % % % scan method % % % % % % finally, the influence of outlier occurrences on the methods performance has been analyzed. accordingly, a new monte carlo simulation similar to the previous one was carried out; the same number of series, with equal numerosity, were generated and some spiked values were introduced at fixed time lags (i = . n, i = . n, i = . n). the monte carlo simulation was replicated five times, varying the outlier magnitude  r . the obtained results were analyzed to compute the total order underestimation percentage. the results reported in table iii show that the expert system is the more affected one, as could be expected, inasmuch as it is based on indices being functions of the series and residual variances. vi. series simulation an infinite moving average representation [ ] is here adopted with terms. the implemented procedure has been compared against the widely used splus software and results have shown the statistical equivalence of generated series. with reference to the series inverse transformation, the following is noteworthy; when the hydrologist preferred box- cox transformation is considered, depending on bc, the generated series can contain data with no physical correspondence. actually, if bc= , an exponential inverse transformation is required, making the simulated peak values significantly increase and so possibly involving maximum wave-height of m, value that actually have never been observed in the central mediterranean sea [ ] or by any ocean buoy all over the world [ ]. moreover, if bc  , the transformed series could have negative values, again with no physical sense. to overcome in some extends these problems, a method to remove negative value is applied retaining the original occurrence of calms (generally defined as sea states having hmo . m); namely, a constant quantity is added to the generated series so that the original number of calms nc is equal to the number of series elements xt . . afterwards, the loess-smoothing procedure is used to trim off the peak values. these shortcomings are eliminated by using plet. vii. application to observed wave data the analysed time series of significant wave-height was recorded by the ron directional wave-buoy located one mile offshore the alghero coast – sardinia (italy), at a depth of about m. the analysed wave record was observed from july to december , with a time interval of three hours, resulting in a series numerosity of observations. for a deeper description of both the italian data buoy network (ron), managed by ispra – oceanographical service, and the measured data see respectively [ ] and [ ]. the missing values are , equal to . % of the data. the frequency distribution of the gap-length showed that almost all gaps cover less than one day and that many of them (equal to the % of gaps) can be simply recovered by neighbouring interpolation. the remaining % of missing data have been recovered by ( ) with  - . , a = - . , b = - . the time series shows no significant daily non-stationarity but there is a clear seasonal component (fig. ) having periodicity of about three months ( hours), which were identified and removed according to the herein illustrated procedure. fig. shows that the seasonal component is markedly affecting the mean value of the observed time series, whereas its standard deviation is nearly invariant within the averaged year. it is noteworthy that the application of the stl robust method give seasonal components much more stable than those computed by the classical averaging method as well as more detailed than those estimated by the fourier representation (fig. ). lag k [hh] k [-] fig. . autocorrelation function of the observed time series at alghero. - . - . - . - . . . . . . . . - . - . - . - . - . - . . . . . . mean component - loss smoothed mean component year fft mean component std component - loss smoothed std component year fft std component - . - . - . - . . . . . . . . - . - . - . - . - . - . . . . . . mean component - loss smoothed mean component year fft mean component std component - loss smoothed std component year fft std component - . - . - . - . . . . . . . . - . - . - . - . - . - . . . . . . mean component - loss smoothed mean component year fft mean component std component - loss smoothed std component year fft std component - . - . - . - . . . . . . . . - . - . - . - . - . - . . . . . . mean component - loss smoothed mean component year fft mean component std component - loss smoothed std component year fft std component  [m] t [dd]  [m] fig. . seasonal components (mean and standard deviation) of the observed time series at alghero estimated by the classical hydrological method, by low order fourier transform method and by the stl robust method (here, the window amplitude used by the loess smoothing is equal to steps, which is nearly equal to one month). time scale ranges from july to june. international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- lag k [hh] k [-] fig. . autocorrelation function of the detrended, deseasonalized series at alghero. f) [-] ) [-] fig. . cumulate frequency distributions of the observed time series as well as those (overlapped) of the standardized series by both box-cox and probability level-equivalence transformations, as a function of the standard variable (ζ). fig. shows the autocorrelation of the detrended time series giving clear evidence of the performance of the applied method. fig. also shows a small residual fluctuation of the autocorrelation with a periodicity of about half a week. no physical causality could be disclosed in this cycling; accordingly, it has been considered tolerable and no more effort to remove it from the detrended series has been done. the time series turned to be non-gaussian as well, and both ( ) and ( ) have been applied to recover the time series normality; the obtained results are shown in fig. . both transformations have performed accurately and efficiently by turning the non-gaussian time series into a gaussian distributed one; nonetheless, the inverse box-cox transformation has showed to be failing by recovering all the storm parametric characteristics when the reverse task of generation is considered. namely, several attempts were carried out by varying the box-cox exponent value bc in the range between and , with a step of . , as well as the width of the loess-smoothing window (see fig. where some of the achieved results are reported). unfortunately, none of the tested combinations gave fully satisfying results by obtaining a simultaneous reasonable agreement between non- exceedance cumulate frequency distributions of both storm duration and its peak-value wave-height. actually, fixing bc = . gives a good agreement between the duration cumulate frequencies but produces too many sea state with unreal giant waves. conversely, setting bc = . gives a fairly good agreement between the storm peak-value wave-height distributions but produces overestimated durations, with sea storminess greatly increased; actually, sea states with hmo over the m threshold persist % more than the observed one. hmo [m] f(hmo) [-]  [hh] f() [-] fig. . comparison between the non-exceedance cumulate frequency distributions of the storm duration (up) and its peak-value wave-height (down), computed from the observed and simulated time series using the box-cox transformation, with different λbc and different width of loess window. with the aim to improve the generation task, the probability equivalence transformation was adopted. accordingly, a probability law should be chosen to properly model the significant wave-height distribution. to this aim, several different probability laws were considered in literature, mainly focused on the distribution upper-tails (see, among others [ ], [ ], [ ]). thought the efforts made, the achieved results do not give any clear evidence of a true distribution [ ]. generally, it is considered that gev type iii describes better the upper tail, at the cost of larger deviations for small hmo values, while the log-normal distribution fits better the distribution mode. thus, gev seems more appropriate for extreme-value analysis (see [ ]), while the lognormal distribution seems more suitable for moderate- value analysis (e.g. fatigue-life analysis, estimation of the wave-energy resource, operativeness analysis, etc.). here, several probability functions were tested and fitted to the observed data by using the complex modified method to minimize the overall least square error. the goodness of fit was verified by using the kolmogorov-smirnov test. the distribution functions having higher k-s confidence level were: gev type iii, three parameters lognormal, four parameters power-lognormal, beta, gamma,  and f (see [ ] for distribution details). each of the above functions was then used in the series simulation, but only the gev type iii and the beta distributions gave fully satisfying results. namely, the lognormal and power-lognormal distributions involve exponential transformation and therefore present drawbacks international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- similar to the box-cox transformation with too many sea state with unreal giant waves. hmo [m] f(hmo) [-]  [hh] f() [-] fig. . comparison between the non-exceedance cumulate frequency distributions of storm duration (up) and its peak-value wave-height (down), computed from the observed and simulated time series using plet with beta and gev iii wave-height theoretical probability distributions. lag k [dd] k [-] k [dd/hh] k [-] fig. . comparison between the non-exceedance cumulate frequency distributions of storm duration (up) and its peak-value wave-height (down), computed from the observed and simulated time series using plet with beta and gev iii wave-height theoretical probability distributions. conversely, the gamma,  and f distributions gave slightly biased distribution lower tails, giving rise to storm persistence too long. accordingly, only the results obtained implementing the gev and beta distributions are reported in fig. ; the attained improvements are evident inasmuch as generated series comply both frequency distributions of storms duration and wave-height peak-value quite well. finally, the beta distribution was chosen for the generation task at alghero owing to its slightly better performance in replicating the duration of the storm with extreme peak-value wave-height (upper tail of the frequency distribution). the parameters of the best fitting beta distribution resulted a= . , b= . , k = . , k = . . the optimal configuration of the linear parametric model at alghero resulted in a second order auto regressive one, with parameters equal to  = . ,  =- . and residual variance r = . . hmo [m] f(hmo) [-] f(hmo) [-] fig. . comparison between the non-exceedance cumulate frequency distributions of hmo observed and simulated by the ar( ) model. hmo [m] f(hmo )[-] hmo [m] f(hmo )[-] fig. . comparison between the non-exceedance cumulate frequency distributions of storm duration (up) and its peak-value wave-height (down), computed from the observed and simulated time series using plet with beta and gev iii wave-height theoretical probability distributions. hmo [m]  [hh] hmo [m] fig. . comparison between the non-exceedance cumulate frequency distributions of storm duration (up) and its peak-value wave-height (down), computed from the observed and simulated time series using plet with beta and gev iii wave-height theoretical probability distributions. fig. shows the autocorrelation functions of the generated and observed series, revealing a nice matching for lags less or equal to a week as well as a nearly perfect seasonal trend. figs. and respectively show the not-exceedance and occurrence frequencies of the significant wave-height. the achieved results are quite gratifying, in that distributions are in very nice agreement starting from values greater than . m, value generally assumed as the lower threshold below which data are discarded both in extreme and climatic analysis. in addition, the over-threshold persistence curves are pretty overlapped (fig. ); it is relevant to stress that the persistence curves almost exactly overlap for hmo > m, given that many international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- engineering activities are limited or broken down by the occurrence of such sea states. table iv. statistical summary of both the series observed at alghero and the ar( ) simulated one. m in st q u . m e a n m e a n lo w e r c o n f. m e a n u p p e r c o n f. s td e r r m e a n m e d ia n r d q u . m a x v a r ia n c e s td d e v . s k e w n e ss k u r to sis observed . . . . . . . . . . . . . simulated . . . . . . . . . . . . . f(hp) [-] f() [-] hp [m]  [gg] fig. . comparison between the occurrence frequencies of storm-peak significant wave-height (hp - on the left) and storm duration ( - on the right) for the observed time series and the simulated one. all these fine agreements reflect themselves into the descriptive statistics of the observed and simulated series (see table iv). the only dissonant note is the mismatch between observed and simulated data kurtosis in table iv, which points out a greater peakedness of observed frequency mode (fig. ). taking all these features in mind makes possible to say that the correlation structure of the observed data is very well reproduced into the synthetic time series. by considering the derived dataset of the storm features, characterized by the peak-value of the significant wave-height and by its duration over the threshold of m, the agreement is slightly less gratifying but still suitable. namely, fig. shows the occurrence frequencies of the storm duration and peak-value. the achieved results indicate that the simulated mild storms, characterized by a peak-value of the significant wave-height in the range of ÷ m, are more frequent than the observed ones. on the contrary, the simulated violent storms, characterized by a peak-value of the significant wave-height in the range of ÷ m, are fairly less frequent than the observed one. for the extreme storms, characterized by a peak-value of the significant wave-height greater than m, both simulated and observed ones exhibit nearly the same frequencies. furthermore, the simulated storms show a little bit shorter duration. actually, results indicate that the simulated short storms, characterized by a duration less or equal to days, are more frequent than the observed ones. on the contrary, the simulated persistent storms, characterized by a duration in the range of ÷ gg, are fairly less frequent than the observed one. finally, the very long storms, characterized by a duration greater than gg, show nearly the same frequencies for both the simulated and observed series. . . . . . . . . . . . . . . simulated . . . . . . . . . . . . . . observed f(hp,) [-] hp [m]  [gg] f(hp,) [-] hp [m]  [gg] fig. . comparison between bivariate occurrence frequencies of storm- peak significant wave-height (hp) and storm duration () for the observed series (on the right) and the simulated one (on the left). hmo [m]  [hh] hmo [m]  [hh] fig. . evolutions of real and simulated sample storms for severe (left) and mild (right) sea-state conditions. these trends reflect themselves into the bivariate density of occurrence frequency for the simulated series (fig. ), which appears more peaked near the origin. fig. states also that the simulated rare storm events (small occurrence frequency) last shorter than the observed one, independently from the peak wave-height. the described deviations can be explained considering the evolution of sample storms for both the observed and generates series, in mild and severe conditions, reported in fig. ; as shown, the storm decreasing-tails decay slightly faster for the generated events than for the observed ones. moreover, the decay stops when generated wave-heights became very small while the observed ones show a more persistent behaviour; accordingly, a possible successive event reaches faster the sea-storm censoring-threshold of m. these elements concur to give a slightly greater duration of the observed storms. viii. conclusion this paper describes an improved methodology for the analysis, missing-value completion and simulation of an incomplete, non-stationary and non-gaussian time series of wave significant height. the method analyses a finite-length time series to identify an arma model, which can be used to recover missing values, forecast short-term wave evolution or to generate arbitrarily long sequences of wave data, preserving the primary statistical properties of original dataset, including persistence over threshold, autocorrelation, non-gaussian distribution and seasonality. three main improvements to the general arma fitting procedure are introduced: a robust estimation of the seasonal components, and accurate method to compute model parameters along with an automatic expert system for the model optimal-configuration selection. these implementations international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- are effectively verified to be fine improvements. namely, stl method computes very regular as well as detailed seasonal components; in addition, the whittle’s approximation coupled with the complex-modified optimization-procedure give parameter-estimates with lower variance and unbiased mean when different series drawn out from the same process are analysed. finally, if the observed series is unaffected by significant outliers, the expert system is able to automatically and properly identify the true model with rates very close to % and, generally, slightly overestimating the model total order and correctly identifying the right prevalence of the ar and ma model parts. the proficiency of the herein proposed methodology is demonstrated in this paper through comparisons of simulated series with data observed by the directional buoy of ron, located offshore alghero coast – sardinia, italy. the achieved results point out that statistical properties of the observed and simulated time series are almost nearly equivalent; this nice agreement embraces winter and summer seasonal patterns, sea state sequencing, over-threshold persistence, occurrence and cumulate frequency distribution of significant wave-height as well as both the cumulate frequency distribution of the storm duration and its wave-height peak-value. accordingly, the described arma-modelling procedure is an efficient tool in representing the wave-height climate. its straightforward application is accordingly associated to the comprehension of sea state conditions, which is of central importance for many offshore and nearshore activities. actually, estimates of risk for critical scenarios are often defined as some over-threshold responses of complex and interlaced systems; monte carlo can therefore be the only way to derive the probabilities of interest. accordingly, even if there is a huge amount of data collected on ocean waves, which is jet geographically sparse and time limited, arma models can be very helpful, being able to provide large database of observed statistically-equivalent information. moreover, taking into consideration the modern engineering-area of wave-energy conversion, a fresh and promising application of linear model is delineated. actually, according with [ ], a real-time control of converters is required to approach the optimal efficiency of wave-energy extraction; to this aim the knowledge of future incident wave elevation is mandatory. treating wave surface fluctuations as a time series and applying an arma model, makes possible to predict incoming wave elevation only from its past history. results achieved on real data from galway bay and pico island showed the proficiency of a linear model to render a very accurate prediction of the incoming swell waves for a lag up to two wave periods. the herein presented methodology can be promptly adapted to wave elevation time series excluding the seasonal components estimation. references [ ] p.td. spanos, “arma algorithms for ocean modelling”. trans. asme, j. energy res. tech., vol. , pp. - , . [ ] r.j. sobey, “correlation between individual waves in a real sea state”, coastal eng., vol. , pp. - , . [ ] c. guedes soares, a.m. ferreira, c. cunha, “linear models of the time series of significant wave height in the portuguese coast”, coastal eng., vol. , pp. - , . [ ] c.n. stefanakos, g.a. athanassoulis, “a unified methodology for the analysis, completion and simulation of nonstationary time series with missing values, with application to wave data”, app. ocean res., vol. , pp. - , . [ ] p.c. ho, j.z. yim, “a study of the data transferability between two wave-measuring stations”, coastal eng., vol. , pp. - , . [ ] s. grimaldi, “linear parametric models applied to daily hydrological series”. journal of hydrologic engineering, vol. , pp. - , . [ ] g.e.p. box, g.m. jenkins, “time series analysis: forecasting and control”, holden-day, san francisco, . [ ] j.d. salas, “analysis and modelling of hydrologic time series”, in handbook of hydrology, maidment editor, eds. mcgraw-hill, . [ ] g.a. athanassoulis, c.n. stefanakos, “a nonstationary stochastic model for long-term time series of significant wave height”, j. geoph. res., vol. , pp. - , . [ ] g.e.p. box, d.r. cox, “an analysis of transformation”, j. roy. stat. soc., ser b., vol. , pp. - , . [ ] c. cunha, c. guedes soares, “on the choice of data transformation for modelling time series of significant wave height”, ocean eng., vol. , pp. - , . [ ] j. beran, “statistics for long-memory processes”, chapman and hall, new york, . [ ] s.m. kay, s.l. marple, “spectral analysis – a modern perspective”, proc. of the ieee, vol. , n° , pp. - , . [ ] m.s. taqqu, v. teverovsky, “robustness of whittle-type estimators for time series with long-range dependence”, stochastic models, vol. , n° , pp. - , . [ ] m.j. box, “a new method of constrained optimization and a comparison with other methods”. computer journal, vol. ( ), pp. - , . [ ] p.e. gill, w. murray, “numerical methods for constrained optimization”. academic press. . [ ] r. piscopia, “on the optimal fitting of a ten-parameter model to observed wave spectra”. isope th conf., vol. iii, pp. - , . [ ] j.a. ferreira, c. guedes soares, “modelling the long-term distribution of significant wave height with the beta and gamma models”, ocean eng., vol. , nº , pp. - , . [ ] g.m. ljung, g.e.p. box, “on a measure of lack of fit in time series models”, biometrika, vol. , pp. - , . [ ] w.s. chan, “a comparison of some pattern identification methods for order determination of mixed arma models”, statistics & probability letters, vol. , pp. - , . [ ] p.j. brockwell, r.a. davis, “introduction to time series and forecasting”. springer verlag, new york, . [ ] t.w. anderson, “the statistical analysis of time series”. john wiley & sons, new york, . [ ] r. piscopia, r. inghilesi, s. corsini, l. franco, . “italian wave atlas”, ed. univ. roma iii, us lib. congr. g . .c , pp. , . [ ] p. petrova, z. cherneva, c. guedes soares, “distribution of crest heights in sea states with abnormal waves”. appl. ocean res., vol. , pp. - , . [ ] m. bencivenga, g. nardone, f. ruggiero, d. calore, “the italian data buoy network (ron)”, proc. advances in fluid mechanics ix, ed. wit, pp. - , . [ ] r. piscopia, r. inghilesi, a. panizzo, s. corsini, l. franco, “analysis of - year wave measurements by the italian wave network”, proc. th icce (cardiff, uk), pp. - , . [ ] m. isaacson, n.g. mackenzie, “long term distribution of ocean waves: a review”, j. waterways, port, coastal and ocean eng., vol. , pp. - , . international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- [ ] y. goda, k. konube, “distribution function fitting for storm wave data”, proc. nd icce conf., vol. , pp. - , . [ ] c. guedes soares, m. scotto, “modelling uncertainty in long- term predictions of significant wave height”, ocean eng., vol. , pp. - , . [ ] m. evans, n. hastings, b. peacock, “statistical distributions”, rd ed., john wiley and sons, . [ ] f. fusco, j.v. ringwood, “short-term wave forecasting for real-time control of wave energy converters”, sustainable energy, ieee transactions, vol. , n° , pp. – , . international journal of engineering research & technology (ijert) issn: - www.ijert.orgijertv is (this work is licensed under a creative commons attribution . international license.) vol. issue , september- llllllllllllllllllllllllllll~llllllllllll *’ ; i-- ’ : .. ‘\t\. ,. : ” ’ i , ‘., ;. marketing planning and expert systems : an epistemology o f practice by dr. m a l c o l m h.b. m c d o n a l d p r o f e s s o r o f marketing planning and director o f the cranfield marketing planning research centre, cranfield s c h o o l o f management, cranfield, bedford m k o a l tel: marketing planning and expert systems : an epistemology of practice abstract after nearly a quarter of a century, artificial intelligence, in spite of all its promise, has made virtually no progress in the domain of marketing, and whilst most interested parties view them as a potentially powerful way of beating the competition, there are few products and no on-line systems available. this paper explores why progress has been so slow in the domain of marketing and describes the experience and progress of a group of major british multinational companies who have joined forces to produce an expert marketing planning system, exmar, with the author of this paper as principal expert. a number of conclusions are drawn, but one of the main ones is that the development of exmnz shows that it is possible to use expert system methodologies to build support systems in complex areas of marketing management, especially if the domain is well defined, has a large number of factors to be considered, and relevant expert knowledge is available. also expert systems are shown as being useful in helping both academics and practitioners to structure, validate and use marketing knowledge and to better understand the interrelationships between the elements of marketing. in particular, it forces managers to think deeply and in a structured way about the issues that need to be considered in developing a strategic marketing plan. marketing planning and expert systems : an epistemology of practice just imagine what would happen to a major industrial company’s profitability if, instead of expert marketing knowledge being hoarded in the heads of an elite but small number of very experienced and successful marketing managers, &l of the company’s worldwide marketing decisions were being made using this expertise. imagine what would happen to a bank’s profitability if &! the decisions were being made by its very best bankers. imagine what would happen to a unit trust company if & the investment decisions were being made by their very best experts. after nearly a quarter of a century of expert systems, dreams such as this now seem possible. but there is still a long way to go, and many formidable technical and methodological obstacles still remain to be overcome. a surprising fact about expert systems is that although they have inspired a number of new programming languages and powerful new computer architectures, they have made virtually no progress in the domain of marketing, and whilst most interested parties view them as a potentially powerful way of beating the competition, there are few products and no on-line systems available . because artificial intelligence has become the latest buzzword, many software houses are hyping up their old software in advertisements, but most of these can be discounted as irrelevant in the real world of expert systems . the principal reasons for this lack of progress centre around the technical problems associated with getting computers to mimic experts and the costs involved. - - there are no shortcuts to building good expert systems. it takes a considerable amount of skill, patience and several years of effort to develop an expert system in a n e w area and get it into the field . the purpose of this paper is: . to explore why progress has been so slow in the domain of marketing and to evaluate the impact that expert systems are likely to have on marketing management. consequently, technical issues are discussed only briefly. for a full technical explanation of artificial intelligence and expert systems, readers should refer to the marketing science institute paper on expert systems in marketing . . to discuss the experience and progress of a group of major companies w h o have joined forces to produce an expert marketing planning system, exmar, with the author as the principal expert. w h a t a r e e x p e r t s y s t e m s ? expert systems is a branch of what is known as artificial intelligence, which is a loosely grouped activity in which a number of researchers of varying backgrounds have done s o m e research since the mid s. but artificial intelligence is still not tightly defined. according to horwitt “artificial intelligence is one of the most misunderstood concepts of our time, and little wonder. the fact that very few real- world ai applications exist only serves to feed our wildest sci-fi fantasies. o n e of ai’s major effects, however, has been the spawning of four critical areas of business computer applications research : natural languages; robotics; visualisation systems; and expert systems.” conventional computing deals with simple and unambiguous facts with existing packages being little more than moronic number crunchers. most software is written in the form of an algorithm, which is a list of c o m m a n d s for the computer to carry out in the order prescribed. it uses data held in a separate file, which is stored in a particular way. thus, software is data plus algorithm and is useful for boring, repetitive, numerical tasks. the largest selling software has been spreadsheets and word processing packages. database management was developed from this. however, managers handle more than words and numbers. they are concerned about knowledge, which is information interpreted for a particular application. the british computer society definition of an ‘expert system is: “t h e embodiment within a computer of a knowledge based component, from a n expert skill, in such a form that the system can offer intelligent advice or take a n intelligent decision about a processing function. a desirable additional characteristic, which many would consider fundamental, is the capability of the system, o n demand, to justify its o w n line of reasoning in a m a n n e r directly attributable to the enquirer. t h e style adopted to attain these characteristics is rule-based programming.” put more simply, expert systems capture not only the knowledge of a h u m a n expert, but also the rules he uses to reach his conclusions. this knowledge is then m a d e available to others by means of a computer program. the two main components of &r expert system are: the knowledge base the inference engine - - the rules used by an expert and his knowledge and experience about a certain domain are interrogated and the captured knowledge becomes the knowledge base, which is the heart of the system. the inference engine accesses the knowledge base, makes the necessary connections, draws conclusions, and generates the answers. the general reasoning strategies are separated from the knowledge base so as to allow the system to use knowledge in a variety of ways, requesting additional information if required to solve a particular problem and explaining the reasoning behind its questions and recommendations by reporting the rules and facts used. since the knowledge base and inference engine are separate, an inference engine can be bought to be used in association with other data bases. this is called a shell. a n expert system will usually have the following characteristics: it will relate to one area of expertise or knowledge rather than to a set of data it will be restricted to a particular topic it will have collected the rules (heuristics) and knowledge of an expert it will have an inference engine it will be capable of extension it will be able to cope with uncertainty it will give advice - - it will explain its reasoning. to summarise, the differences between traditional packages and expert systems are . as follows . traditional packanez handles data uses algorithms goes through repetitive processes based on large data bases exdert svstems handles knowledge uses heuristics goes through inferential processes based on knowledge bases w h y h a s p r o g r e s s b e e n s o s l o w i n t h e d o m a i n o f m a r k e t i n g ? during the os, attention w a s focussed on specific problem-solving applications in scientific fields. m a n y successful expert systems have been built, including mycin for diagnosing infectious diseases’, and p r o s p e c t o r , a system for evaluating geographical locations for possible mineral deposits . management problems, however, do not lend themselves to quite the s a m e precise logic as scientific problems. people do not solve most of life’s problems by mathematical means, but rather by experience, knowledge and intuition. marketing problems are dealt with in the s a m e way, as most of them are logical rather than mathematical, and problem-solving knowledge, whilst available, is incomplete. decision-support systems and the like use hard facts and static formulae which, given the correct data, provide correct answers. they belong more naturally to the logical, black-or-white, right-or-wrong world of computers. but managers in the world of marketing deal with uncertainties and often with vague concepts. decisions invariably are built on a set of “rules”, or heuristics, that reflect the expert’s o w n knowledge and experience about the problem in question. these “rules” are hard to nail down and quantify, because the expert’s experience enables him to think in terms of shades of grey, “more or less”, and “approximately”. such fuzzy reasoning is commonly used by h u m a n beings to find a path through situations that are too complex and amorphous for the h u m a n mind to handle in a totally conscious, rational, scientific way. most people would acknowledge that in virtually any walk of life, the true expert has built up his expertise largely from experience and an intuitive grasp of problem- solving in the real world, something which is often referred to ‘as the “university of life”. indeed, m a n y of the world’s leading business people acknowledge that they o w e their success not to formal business education and text books, but to their o w n experience, flair and intuitive good judgement. donald schon describes this phenomenon as follows: “competent practitioners usually know more than they can say. they exhibit a kind of knowing-in-practice, most of which is tacit”. h e cites an investment banker, w h o makes his decisions based on to per cent instinct, and only to per cent calculable rules. this “gut feel” was a major asset to the bank in question. his point is that artistry is not reducible to discernible routines. h e describes scientific rigour as “describable, testable, replicable techniques derived from scientific research, based on knowledge that is testable, consensual, cumulative and convergent”, but then goes on to argue that m u c h of what passes for scientific - - management is irrelevant because business problems do not c o m e well formed. certainly, most marketing problems are messy and indeterminate and successful practitioners m a k e judgements using criteria and rules which are difficult to define. m a n y academics would decry this as a lack of rigour, and in so doing exclude as non-rigorous m u c h of what successful practitioners actually do. the following quotation from schon neatly sums up the problems facing not only teachers and researchers of marketing, but, more importantly, the initiators of expert marketing systems : “in the varied topography of professional practice, there is a high, hard ground which overlooks a swamp. o n the high ground, manageable problems lend themselves to solution through the use of research-based theory a n d technique. in the swampy lowlands, problems are messy a n d confused a n d incapable of technical solution. t h e irony of the situation is that the problems of the high ground tend to b e relatively unimportant to society at large, however great their technical interest may be, while in . the swamp lie the problems of greatest h u m a n concern.” the problem to be addressed by expert systems in the marketing domain, then, centres around h o w to take account of the intuitive artistry displayed by experts in situations of complexity and uncertainty in a way that is describable and susceptible to a kind of rigour that falls outside the boundaries of technical rationality. the question, then, is h o w an e&temology of practice can be captured and represented in an expert system. for an expert system to mimic an expert, it needs to be able to deal with the uncertainties, complexities, and vague concepts that h u m a n beings deal with routinely, even though such “rules” are neither simple nor straightforward. for example, a simple rule for a marketing manager might be: “if the market is growing, increase promotional expenditure”. this would appear to be easy for a h u m a n being to understand, but in reality words like “market”, “growing”, “increase” and “promotional expenditure” are open to m a n y different interpretations, as indeed is the whole lexicon of marketing. o n e way of dealing with this problem is the development of fuzzy sets. a “growing market”, for example, is a fuzzy set in the sense that its meaning can vary from situation to situation. fuzzy numbers approximate the response figures from marketing experts and these numbers are then loaded into, for example, sales projections and promotion analyses. the foundation of any expert system is the knowledge base, which can be extracted from one or more experts in a particular field. the expertise is usually stored in the form of rules of thumb (heuristics), which are, typically “if then” statements. for example, if a is true, then b is true; or if x is true, do y. given an initial set of circumstances, the system can m a p out a set of contingencies and further -- contingencies. a heuristic differs from an algorithm in that it does not give a correct answer, nor does it guarantee results. it merely suggests a general direction that is more or less likely to prove more useful than another direction. a n example of a heuristic in chess might be: “if a player stays in control of the centre of the board, he is more likely to win”. in marketing, a heuristic might be: “if the market is growing and if you have appropriate business strengths, then an appropriate marketing objective would be to grow market share”. a system of interlinking heuristics in the form of a decision tree is one way of representing knowledge. these are sometimes “backwards inferencing” and sometimes “forward inferencing”. backwards inferencing starts with an objective and tries different combinations of rules and/or actions until it is reached. forward inferencing reasons from initial information until it reaches useful conclusions. this can give rise to what is termed “combinatorial explosion”, which can be avoided by pruning and the use of heuristics which are correct most of the time. this gives probable solutions to less rigorously defined problems that are too complex to be dealt with algorithmically. to date, however, no one has seriously tackled the world of marketing with expert systems other than the msi m x x d ~ system developed to advise on advertising design. after considering a variety of consumer and environmental factors, advertisers use a combination of empirical research, communication theory, and rules of thumb, to select communication objectives and select appropriate creative approaches. the authors themselves list a number of weaknesses in a d c a d , but conclude: “as one advertising executive put it: “it helps us to think a little deeper about the issues w e have to consider in developing ads that are both strategically a n d executionally sound”. another interesting and relevant conclusion was that most managers, when asked, said they would like to m a k e use of existing theoretical and empirical knowledge of marketing when making decisions. however, few actually did use such knowledge. expert systems can bridge this gap by structuring, validating and disseminating marketing knowledge, whilst at a theoretical level, they challenge their creators to understand and critically evaluate the elements of marketing knowledge and their interrelationships. a c a s e history o f t h e d e v e l o p m e n t o f a n e x p e r t s y s t e m in m a r k e t i n g planning during the os, japanese activity in the field of expert systems prompted the e e c to give birth to the esprit programme in an attempt to integrate european efforts. this in turn led to the dti sponsored alvi programmes. a n outcrop of these is a n e w dti-sponsored club by the n a m e of exmar, which set out in to produce an expert system in the domain of marketing planning, inviting the author of this paper to be the principal expert. the ten founder m e m b e r companies include s o m e of britain’s biggest and most successful multinational corporations spanning capital goods, industrial goods, consumer goods, and service industries. after almost two years of work and an expenditure of over f) million, all there is to show is a demonstrator model on a xerox workstation which exemplifies the scope of the expert system using a case study specially written for the club by the author of this paper. the purpose of this part of this paper is the explain h o w e x m a r has developed, what obstacles were encountered along the way, h o w these were overcome, and what problems still remain to be solved before a commercially usable p c based system can be m a d e available. the first point to be m a d e is that expert systems do play a vital role in the accumulation, synthesis and understanding of the constructs of marketing and their interrelationships. m a n y of the theories, illuminative sights, empirical research findings, models, and experience, are scattered around in books, libraries and inside the heads of both practitioners and academics. they remain, therefore, largely unavailable to most marketing managers, and indeed to most marketing academics. the synthesis of such knowledge in a particular domain into expert systems not only benefits those whose task it is to develop the system, by forcing them to turn their knowledge and expertise into actionable marketing propositions, but also those responsible for marketing decisions by making it available where it is likely to have the greatest impact. p r o b l e m s suitable f o r expert s y s t e m s in deciding whether marketing planning was a sensible domain for the application of expert systems methodology, the m s checklist proves useful. four criteria are provided: are the key relationships in the domain logical rather than arithmetical ? since the decision area is knowledge-intensive, the answer here is “yes”. is the problem domain semi structured rather than structured or unstructured? well-structured problems can use more conventional procedures, but since the marketing planning process is only semi-structured, the answer is “yes”. is knowledge in the domain incomplete ? since marketing planning and all its contextual problems remains one of the most under-researched areas of marketing, and since little has been published about the interrelationships of all the techniques of marketing in systems design, the answer is “yes”. this is in fact the key to the whole project and why it was chosen in the first place by the club members. n will problem solving in the domain require a direct interface between the manager and the computer system ? the intention is to have operational marketing managers using the system for the production of marketing plans, so the answer is “yes”. marketing planning remains one of the last bastions of ignorance in the field of marketing. the benefits of marketing planning are‘ well documented and agreed,l’ yet so complicated is the process of marketing planning, and so confusing are the interrelationships between the tools and techniques of marketing planning , that very few british companies enjoy these benefits, as has been shown by a seminal paper by greenley that reviewed all the major u k empirical research in this area. indeed, there were as many dysfunctional results from the attempts of companies to initiate marketing planning procedures as there were benefits. the whole thrust of the project, then, was to tackle this problem by means of an expert marketing planning system codenamed entar. marketing planning can be defined as a logical sequence and a series of activities leading to the setting of marketing objectives and the formulation of plans for achieving them. the model taken to represent the marketing planning process was the author’s nine stage breakdown , as given in figure later in this paper. - - analysis p h a s e the initial requirements analysis produced a number of interesting problems for the project, which were to s o w the seeds of expensive and time-consuming delay. these problems can be summarised as follows:- (i) it b e c a m e clear that not m a n y of the m e m b e r companies were particularly au fait with the methodology of marketing planning. this led to the problem of setting clear objectives for the project. (ii) the diversity of c o m p a n y industry types, ranging from capital goods to service industries, meant that no subsequent system could possibly be suitable for all circumstances. (iii) problems and subsequent proposed objectives ranged from .“to support a formal planning framework to improve discipline during the planning process” and “to support further understanding of the effects of currency fluctuations” to “to promote discipline in pricing control” for these reasons, it w a s decided to focus on the process of marketing planning itself rather than on any situation-specific system. m e t h o d o l o g y a firm of software consultants w a s appointed project manager and a knowledge based systems house w a s appointed principal contractor. considerable confusion surrounded the proposed delivery system with the result that specifications, such as model, functional requirements, system structure, information requirements, enhancements, consequences, knowledge base specification, validation procedures, and so on were never produced. the systems house began a series of twelve half day interviews with the author of this paper in order to develop the knowledge base. unfortunately, although taped and transcribed, they were largely unfocussed due to the inexperience of the interviewers and little progress was m a d e towards formal modelling of the marketing planning process, in spite of very specific guidance given by the author to the interviewers. the problem, centred around lack of proper project control by the project managers, confused expectations by members of the club based on marketing planning naivety, the inexperience of the knowledge engineers, and the passive role of the domain expert, which was necessary in view of the nature of the project. several attempts on the author’s part to guide the system were brushed aside as politically inexpedient. the result was that the paper outlining the tasks to be performed by the computer system targeted the whole marketing planning process rather than any subset, and because of this breadth, the process to be computerised was not documented in any detail, nor backed up by any substantive models and interrelationships. n e w k n o w l e d g e engineers appointed at this point, the problems began to assume crisis proportions, and the project manager appointed n e w knowledge engineers to take over the feasibility study and the delivery system. - - the n e w contractor set about finding s o m e c o m m o n requirements a m o n g end users in order to outline the domain model, with a boundary definition showing which parts of the model would be tackled by the computer system. they set about establishing the following areas: \ scope constraints organisational impact maintainability extensibility technology time scales risk and cost versus quantifiable benefits for the first time the e x m a r project was beginning to focus on building a system for appropriate problems that were valued, bounded and routine. the following emerged as the final overview of the objectives of e x m a r as agreed by all members of the club. w h a t w i l l e x m a r d o ? e x m a r is intended to be a marketing planner’s assistant. it will guide a user through the marketing planning process, offering advice at key stages, controlling data input and presenting data in various ways so as to assist in the setting of objectives and strategies. - is - the full marketing planning process has nine stages, with various feedback loops, as shown in figure . fbure -the marketind planning p r o c e s s . corporate objectives . marketing audit . s w o t analysis . assumptions a . marketing objectives & strategies ltestimate expected results . identify alternative plans & mixes . programmes . measurement & review the current vision of e x m a r concentrates on stages , and because club members have consistently identified stage (objectives and strategy setting), together with the preceding data collection and analysis, as the main problem areas. corporate objectives and mission statement are taken as the given inputs (from outside the user’s influence) needed to start the process. all relevant data is then collected in a marketing audit phase. this data is then abstracted and analysed in the s w o t phase and relevant assumptions recorded. various methods are then available in the final phase to assist the user to set realistic and consistent objectives, together with coherent strategies to meet them. - - it is anticipated that e x m a r sessions will be highly interactive and iterative, encouraging scenario planning. they should also permit analysis at different levels of detail, from a corporate overview of key business sectors to in- depth studies of individual market segments. further details of h o w members believed e x m a r would actually do this, together with implementation and development constraints, are included as appendix . from this it will be seen that members wanted ibm pc compatibility for hardware, with software amenable to change by programmers not involved in its development, and which would be amenable to extension and add-ons. t h e n e x t s t e p a number of refinements and corrections to the methodologies and interrelationships was n o w necessary before the project could proceed. these were detailed by the author in a separate document, the relevant part of which is reproduced in appendix to this paper. d e v e l o p m e n t o f a d e m o n s t r a t i o n m o d e l s o m e further interviews with the knowledge engineers quickly moved the project towards the production of s o m e deliverables. it was possible, for example, to define those parts of the marketing planning process which seemed the most likely candidates for automated support. the agreed primary objective of e x m a r was to provide automated assistance for the marketing planning process, since it had been agreed a m o n g members that in general marketing decisions are taken without sufficient analysis and understanding of the relevant issues. the reason was seen as being a lack of knowledge and understanding of h o w and why the multifarious factors of marketing interact and serve to form the parameters of any business activity. in this real life situation w e see emerging the perfect role for an expert system in marketing planning. d e m o n s t r a t i o n m o d e l all that remained n o w was to produce a model to demonstrate h o w such an expert system would work. for this, the author wrote a special case study based on a multinational company in the bearings industry. the case study contained all the necessary features to demonstrate the scope, methodologies and outputs of the proposed expert system. a detailed report was a necessary prerequisite for producing a live demonstration model. the report actually produced outlines the scope and functional breakdown, the data model, and the technique interrelationships. this report is included as appendix . it is recommended that this should be carefully studied, as it describes the basic model and outlines the technique interrelationships. not included in this paper are other parts of the report relating to technique descriptions, model testing and technical details relating to the demonstration itself. from this it will be seen that the step model shown in figure was m a d e more amenable to computerisation, as shown in figure . a n example of the detail included in one of these stages is shown in figure . the basic data model used and s o m e of the techniques relating to it are shown in figure . - - f u r t h e r r e f i n e m e n t o f s c o p e figure various related areas are outside exmar's scope, on the grounds that, though important, they are peripheral to the central concerns of exhur, and should not be studied in detail in the interests of timely focus. these areas are summarised in the boxes on the diagram below outside the “scoping” dotted line. brief notes on these follow. i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i s e l e c t / d e f i n e b u s i n e s s unit i d e f i n e unit m i s s t o n i i i i i i i i i c o n o u c t audit s u m m a r i s e i i i i s e t o b j e c t i v e s -[set] i i i positloning i i m-- - -------- ----------m- --- --------------------- : . i i p r o d u c e s t r a t e g i c ; m a r k e t i n g p l a n f o r a i i i i b u s i n e s s unit i s e g m e n t a t i o n conduct audit figure c o n d u c t audit .- f a c t o r s .- .b checklist * / a s s e s s critical - s t r e n g t h s s u c c e s s f a c t o r s t a b l e i a n d w e a k n e s s e s i a s s e s s o p p o r t u n i t i e s and- t h r e a t s c l .w of .- .d c h e c k l i s t * o t list a s s e s s m a r k e t -~ a t t r a c t i v e n e s s l ‘+=j a~~tk: czg the objective is to assess the state and prospects of the products and markets already identified. information needed at this point m a y have been collected in advance of the planning process, or it m a y be collected now. - - techniaue interrelationshins data used by techniques figure the diagram below shows the data used as input by some of the techniques modelled. market maet ;;;; mj _-_. sech ia-*- ,.~~~~:~~~ \ \ (actual orpote$ai) oefmmon compe titer) l planning g l name or . nnaa ‘,.l‘illra ’ i l differentiatior/ (l- ) . . / product for l market share’ -. i ‘). .costs --__ i -. score on csfi,- ;-o~a.u&re newlexisti ng . , --s_ price/average priq ’ i ’ i ’ i ’ i ’ , ’ , l relation to sales volume ‘. *. i ‘\ i , -. , -. / .‘- [-+q is sold in to .’ product i i l name: i i l costs’ : d newlexistind (self or compe tr tar) j production of a demonstration model at a packed meeting of the members in december , the demonstrator model was unveiled. its purpose was: to demonstrate how such a system would meet the club’s primary objectives; to provide evidence of the feasibility of building such an expert system in technical terms; to provide a basis for feedback about the systems’s utility. it was developed on a xerox workstation running the interlisp environment to minimise the time required to build the demonstrator and because of interlisp’s power and maturity. the demonstrator provided: guidance and support for the marketing planning process at various stages and help in managing the interactions; variable forms of information presentation and manipulation, such as data forms, diagrams and text. relationships and constraints between information are managed by the system, for example by calculation and iconic cross- references; a free interface which allows the user to take the initiative in determining precisely what he wants to do next, and what he wishes to have displayed to assist his actions. this is done by the provision of a number of means of navigating around the window-based system. - - the demonstrator model was spectacularly successful with club members and clearly illustrated the large amount of iteration that would need to occur in generating a plan. it also gave s o m e indication of the processes of information gathering and debate that would typically have to occur in the real world whilst using the system. conclusion although the actual demonstration model using the case study is not included in this paper, for reasons both of confidentiality and brevity, the e x m a r project has clearly reached a stage of development that demonstrates the value of expert systems in marketing. a number of conclusions can be drawn: ( the development of e x m a r shows that it is possible to use expert systems methodologies to build support systems in complex areas of marketing management, especially if the domain is well defined, has a large number of factors to be considered and relevant expert knowledge is available. (ii) the more complex and amorphous the expertise to be captured, the longer it takes both the expert and the knowledge engineer to reach an acceptable approximation. it is clear that to develop an expert system that is of s o m e practical use requires both time and resources of massive proportions. this is supported by the m s research paper , which concludes: “there are no shortcuts to building a good expert system. it takes a considerable amount of skill, patience, and years of effort to develop an expert system in a n e w area and get it into the field”. - - (iii) expert systems provide a consistency to h u m a n decision making which is valuable, since people tend to forget or ignore knowledge. (iv> e x m a r has generated considerable interest and support a m o n g the major multinational companies that form the club, because it forces them to think deeply and in a structured way about the issues that need to be considered in developing a strategic marketing plan. (vi expert systems are useful in helping both academics and practitioners to structure, validate, and use marketing knowledge and to better understand the interrelationships between the elements of marketing. ( tight project control is vital. this view is supported by mumford . in particular, the following issues need to be considered: ) subject matter - h o w well it is defined ? - is it likely to change during the project’s life ? - can adequate inputs be provided by both experts and.knowledge engineers ? (ii) t h e user - do they understand the likely time of the project ? - do they know exactly what they want ? - are they willing to work constructively to solve problems ? (iii) time - are the project deadlines realistic and achievable ? - - (iv) resources - is the budget sufficient ? - is sufficient skilled human resource available ? - will facilities requirements be catered for ? (v> project management - is the project management strong enough and sufficiently disciplined? (vii) the potential advantages of expert systems are: n consistent advice a secure knowledge bases n making better use of experts n enhanced decision making n improved analysis (viii) the stages in building an expert system are: ( problem identification and definition (ii) the acquisition of relevant knowledge . (iii) the representation of relevant knowledge (iv) the selection of a reasoning approach (vi system selection (vi) prototype development (vii) system refinement and validation. (ix> since we live in an imperfect world, with imperfect problems and imperfect tools, it is unreasonable to expect a perfect expert system until there are perfect experts and perfect technology. on the other hand, if an expert system gives better advice than you would have had without it, it is probably worthwhile. in conclusion, it is unlikely that expert systems will ever be able to give the s a m e value as real h u m a n experts, although clearly they can offer reasonable advice. nor will they guarantee that you m a k e the right decisions. but they can help you gain a proper perspective of the alternatives. in a sense, expert systems will always be a bit like distance learning programmes, which can replace a bad teacher, but never a good one. r e f e r e n c e s . . . . . horwitt e. “exploring expert systems” business computer systems, march, . bryant n. “managing expert systems” john wiley and sons, . buchan b. and shortliffe e. “rule-based expert programs : the mycin experiments of the stanford heuristic programming project, reading, ma., addison- wesley, . . . . . . moutinho l. and paton r. “expert systems : a n e w tool in marketing” quarterly review of marketing, summer, foster e. “artificial intelligence” personal computing, april, cebrzyask “artifical intelligence : the goal is to store an expert’s real knowledge on a disk” marketing new, vol. , no. , rangaswamy a., burke r., wind j. and eliashberg j. “expert systems for marketing” marketing science institute working paper report, nos. - , duda r., gaschnig j. and hart p. “model design in the prospector consultant system for mineral exploration” in “expert systems in the micro- electronic age” ed. michie d., edinburgh: edinburgh university press, ’: schon d. “the crisis of professional knowledge and the pursuit of an epistemology of practice”, research paper for the harvard business school th anniversary colloquim on teaching by the case method, april, mcdonald m. “the theory and practice of marketing planning for industrial goods in international markets”, cranfield institute of technology, phd., leppard j. ‘“marketing planning and corporate culture - a conceptual framework which examines attitudes in the context of marketing planning” cranfield institute of technology, m.phil., greenley g. “a n exposition into empirical research into marketing planning”, journal of marketing management, .l., july, mumford e. “designing computer based systems”, university of wales review business and economics, , appendix h o w will e x m a r achieve its objectives ? firstly the system will prompt the user for information to define the business unit to be analysed. this will include a basic definition, mission statement and top level objectives. the user will then be asked to specify the market segments and products to be analysed (the system uses a simple pareto / rule to help the user to focus on the most important business areas). the result is a comprehensive list of existing and potential product-in-market combinations which can be organised in the form of an ansoff matrix (figure ). figure , nsoff matrix increasing technological n e w n e s s d p r o d u c t s existing potential m existing m a r k e t p r o d u c t a p r e s e n t a t i o n d e v e l o p m e n t r k e t m a r k e t diversifi- s potential e x t e n s i o n cation the system will ask the user to order the products and markets by h o w n e w they are to the business unit’s existing area of operation. - . - at this (audit) stage the ansoff matrix is used to drive the data collection process. it can be used in later stages to a s s e s s strategic direction b y reference to the classification of product/market combinations in each box (eg. market extension versus product development). o n c e a complete list of markets and products has been established the s y s t e m will prompt for k e y information required for the s w o t analysis and later stages. for each market s e g m e n t the user m u s t supply two sets of factors: one to m e a s u r e the attractiveness of the market to the business unit; the other to m e a s u r e h o w a product m a y be evaluated b y that market. t h e s e are k n o w n a s market attractiveness factors and critical s u c c e s s factors ( m a f s and csfs). e a c h product m u s t then be evaluated for each relevant market s e g m e n t b y inputting scores against the c s f s previously specified, to a s s e s s business strength (relative to the competition). this has to be done for both the current position and the forecast position. forecasts are also required of performance level. t h e s w o t stage also requires information on opportunities and threats, in terms of their impact and likelihood. t h e s e can be s u m m a r i s e d in an impact/urgency matrix and referenced at later stages, where the user is reminded of threats of high impact and likelihood w h e n setting strategies. a large a m o u n t of analysis is necessary to support the s w o t stage and e x m a r will be able to a c c e s s other packages for supplementary analyses. a s s u m p t i o n s are input a s part of the forecast and once the s w o t is complete the user can proceed to set objectives and strategies. - a - objective setting is driven by the concept of g a p analysis which portrays the target level of performance and a ‘status quo’ forecast figure. the forecast is obtained from summarising the performance level of all product/markets in the top left hand corner of the ansoff matrix. the user attempts to close the gap by: a) selecting existing product/markets and improving operational performance b) selecting n e w product/markets for inclusion in the portfolio. the key aid for the user in this process is the directional policy matrix (fig. ) which essentially summarises a large amount of the s w o t analysis. figure directional policy matrix business strengths (csf scores) h i g h m e d l o w market attractiveness (maf scores) h i g h m e d size of circle shows performance level this is a very versatile tool which visually displays a large number of the measurable criteria relevant to the selection decision, including performance (size of circle) and potential for improvement (position on the matrix). the user will be able to m o v e circles to more favourable positions on the matrix to represent changes in objectives and such movement is recorded by the system. later, the user will be prompted for strategies to achieve the movement. the result of this stage will be a set of revised product-in-market objectives with associated strategies. at any point the evolving strategy can be evaluated and the system will produce reports and various displays to assist this evaluation. for example, portfolio balance can be evaluated using the directional policy matrix itself, plus other tools including the ansoff and i boston matrices. this implies considerable feedback from strategy to objectives setting and the system will document reasons for changes. this phase is potentially very rich in expertise, and research into ways of capturing this is continuing within the club. w h y d o w e w a n t e x m a r ? the system will provide:- i ) an automated implementation of a rigorous marketing planning process. historically the process has been difficult to implement rigorously. ) a comprehensive statement of the data requirements of the marketing planning process, with particular emphasis on the quantification of previously nebulous concepts such as business strengths. - - ) powerful visual displays of key information. these aid understanding and communication. they also free the user to concentrate on other expertise-rich concepts such as coherence and consistency of strategy. ) an opportunity to build a hierarchical structure of plans, from business overview to detailed product analysis. this will depend on the quality of implementation. the benefits of the above, in terms of the quality of plans (and of the debate during their construction) are similar to those claimed by formalised marketing planning. n o existing software approaches the functionality of that envisaged by exmar. jmplementationconsiderations oreanisation organisationally, e x m a r simply requires the existence of a marketing manager to use the system. naturally such a user will need access to the required data, s o m e of which m a y not be available immediately. o n e of the spin-off benefits of e x m a r m a y be to act as a catalyst to prompt change both organisationally and in the data collected. d e v e l o p m e n t c o n s t r a i n t s the club has consistently specified i b m p c compatibility for hardware and this has not changed. software is a more flexible issue but the following are requirements which should be borne in mind. a) club m e m b e r s are very likely to want to develop and customise their copy of the system. this implies:- ) software which is amenable to change by programmers not involved in its development. ) a preference for software which has a wide user base, particularly a m o n g club members. ) s o m e level of system documentation. b) expertise is likely to be gained in using e x m a r over time, which will generate a need to ‘build-in’ further levels of expertise. thus the software should be amenable to extension of the expert system aspects, implying a rule-based or list processing capability. c) the software should have adequate data communications for access to and from other packages. this requirement also indicates a likely future requirement for a multi-tasking environment. - - d) s o m e club members have been conditioned to expect goldworks to be the chosen software and m a y already have committed themselves to this package. appendix refinements a n d corrections to the methodoloeies a n d interrelationshids outlined in a letter to the club workine bv professor malcolm mcdonald . t h e directional policv matrix ) it’s o k to have nine boxes, which is in any case the w a y it w a s originally conceived. i personally keep it to four because it is conceptually easier and fits more comfortably into what “students” have b e c o m e used to via ansoff, boston, porter, et al. nonetheless, the nine box matrix does provide more options and greater flexibility. c a n i suggest that, rather than confusing users at the construction stage with a nine box matrix, w e only put the lines in after the calculations have been completed. w e must, then, ensure that the dividing points are . (or . ) along each axis. (ii) it is imperative that you do not try to use profit as a measure of circle diameter. take it from me, every time this measure is .used, it distorts the truth. for example, there m a y be a product or market that accounts for, say, per cent of sales value, but per cent f‘“profits”. this would appear as a small circle, so masking its true use of resources. in any case, profit is an accounting notion which depends on an arbitrary allocation of overheads. there is also the tricky question of whether it is products or customers that determine profitability. it is usually the latter, which is rarely catered for in accounting systems. in any case, profit is almost certain to be strongly reflected in the marketing attractiveness criteria. (iii) w e should m a k e it clear that there are m a n y different levels of analysis. this could involve any of the following:- . . - regions (of the world) - countries - areas (of countries) - companies - strategic business units - divisions - product groups often synonymous with markets - products - segments - customers - distributors/agents/wholesalers etc. each one of these can be further sub analysed, if necessary. (iv> w e should ensure that users are m a d e aware of the pitfalls, which are as follows: (a> users must beware of becoming emotionally involved in their o w n interpretation of “attractiveness”, which often leads to them “fiddling” the system to ensure their business comes out in the yoper quadrants. it is clearly illogical (or at least unusual), if everything is seen as highly attractive. in such a case, either all are equally attractive, or the scoring is wrong. the scale - is meant to represent relative attractiveness according to their o w n criteria, so that something that is near to is nothing like as. attractive to the company as something that comes out at, say . . but is does not necessarilv m e a n it is unattractive. to m a k e this effective, perhaps w e should put in a suggestion that users might think in terms of “potential” if they feel (after being given due “warning” of the pitfalls) that the word “attractiveness” might cause problems. o n e other warning. users m be prepared to score , where appropriate. w e might even put in a proposal that, if appropriate, the scale might have negative values to go along with a negative scoring system. this might be appropriate where there are very wide extremes. lb) o n e final point on this. recently, i had a case of a company which was experiencing decline in u its segments. in this case, “attractiveness” hardly seemed like the appropriate description. so, instead w e used “potential”, since s o m e divisions had greater potential for growing sales and profits than others. for example, the “shipping” market was in decline, and the company had a high market share. in the “food” market, on the other hand, (also in decline), the company had a m u c h smaller market share, so the potential for taking market share (and improving profit), was greater, hence it appeared in the upper quadrant. not surprisingly, “food” also appeared on the right of the horizontal axis. the point is that had w e not used this device, everything would have appeared in the bottom part of the matrix. whilst this is obviously a possibility, it would not have been particularly helpful in this somewhat sad case. whether the total picture in this case is acceptable or not is irrelevant. the truth is that this company has diversified into other unrelated business areas that are m u c h more attractive. h a d these newer s b u s been included in the analysis, then clearly even “food” would have appeared as low in attractiveness. (vi a propos the two situations (t- to t. and t. to t+ ) for the vertical axis, i must stress m y strong reservation that his will almost certainly confuse most users, and might even irritate them having to do it twice. nonetheless, it’s perfectly logical, consistent and feasible. what w e must stress, however, is that the first part of the exercise must be for t- to t. and u reflect what has happened historicallv. the m part of the exercise, t. to t+ , is a forecast of attractiveness and must reflect their view of what will happen over the next three years. - . ouantifvinn opportunities a n d threats first of all, let’s consider the following generalised list of macro and micro factors which might be relevant: demographic economic technological macro political legal social/cultural customers competitors distribution channels suppliers potential competitors . a m u c h more detailed checklist will be provided with the actual system. these might be considered to be either qpoortunitia or threats. i suggest w e provide a matrix (similar to the issues matrix) for each, ie. two matrices, one an o p p o r t u n i t y matrix, the other a threatmatrix. w e could m a k e it work as follows:- (a) list threats (no more than ten) (b) probability of occurrence (within t. to t+ ) (. to . ) (c) impact on the organisation (score to ) the matrix would look as follows: impact . i . - - ---.,---- - ----------_ ( . i i probability i op i o c c u r r e n c e - , ’ i i . it can be seen here that threat , say something specific to do with c a p ( c o m m o n agricultural policy), will have a big impact on the organisation (score ), and that there is a high probability that it will happen (probability . ). all threats can be plotted using this methodology, which would need s o m e guidelines similar to those provided in m y issues priority matrix. the whole process can then be repeated for qrmortunities. . strategies arising out of directional poiicv matrix a n a h + if w e get users to predict the scores of critical success factors, then clearly they will need to convert these into strategies. w e must be careful not to lead them too m u c h “by the nose”, and i suggest w e don’t need to go beyond the overall guidelines suggested by the shell directional policy matrix. clearly, these must be converted into x ps jargon by the user, but it would be a gigantic task to attempt to list all possible combinations of marketing mix strategies. also, the gverall objectives for each product/market should be consistent with the guidelines suggested by the porter matrix a n d by life cyc!e analysis. l. g i. e i’l s”$ ;*a. . y. li i - .n - . market life cvcle ‘. w e must be careful here. there is no general recognition iri’*’ ” i’ i”- theory of anything called a “market life cycle”. i have sent under separate cover a detailed explanation of what i mean. but yes, of course the guidelines can be included to help users select appropriate strategies. . a p p e n d i x scode. functional breakdown. data model a n d techniaue interrelations of exmar scone a n d functional breakdown this section defines the functions performed during the relevant stages of the marketing planning process, and in s o m e cases breaks down the functions into simpler functions. the functions are related to techniques and methods used in carrying out the function, and to deliverables that form part of the marketing plan, as defined by mcdonald. the top level breakdown is used to refine further the scope beyond the definition contained in rl initial findings report. n o t a t i o n o f f u n c t i o n a l b r e a k d o w n d i a g r a m s the diagram beiow summarises the notation used in the functional breakdown diagrams. function boxes represent tasks to be performed as a step towards production of a marketing plan. technique/method boxes have icons that illustrate the style of representation used by the techniques or method. d f ”n c t lon j .~~~l.-:“i”““““i:-- i /,,,i . i “m a y mvolve feedback co” “deliverable” t o d level breakdown and scodinp scoping defined by initial findings report the initial findings report defined the scope of exmar phase as being the marketing audit, s w o t analysis and objectives and strategies stages of mcdonald’s stage breakdown of the marketing planning process. this is taken as the starting point for this section. the diagram below summaries this. il. coy orate objectives --- + ,,-----------\ \ , \ ‘\, . mafketing audit ( ‘,, the marketing planning process ‘ , . s w y t analysis ( ‘,, \ \ \ \ \ \ \ , . assumptions i i \ \ \ \ \ \ \ \ \ , \ \ i . ‘\ . marketing objectives and strategies ‘n \ \ \ l.,,-e-. ----a------------ a input to the model. . estimate expected results \ ) . identify alternative plans and mixes . programmes . measurement and review objectives are set subject to certain assumptions: other than this, little formalism has yet emerged with regard to assumptions. the setting of corporate objectives is outside phase scope. so any information from the corporative objectives required is regarded as an t o p l e v e l f u n c t i o n a l b r e a k d o w n it is useful to produce a slightly differing top level breakdown than that contained in mcdonald’s g-stage diagram. this is given below. explanatory notes follow. p r o d u c e s t r a t e g i c m a r k e t i n g p l a n f o r a b u s i n e s s unit s e l e c t / d e f i n e b u s i n e s s unit d e f i n e unit m i s s i o n c o n d u c t audit -j s u m m a r i s e - s e t s t r a t e g y - - produce strategic marketing plan for a business unit this describes the task being modelled. .it is strategic because it involves ignoring some details to aid clearer thinking about the important parts of a business. it is for a business unit because this process can be carried out at any level of an organisation, or for a subset of the business that crosses organisational boundaries. select/define business unit identify which area of the business the marketing plan is for. define unit mission define what the business unit is in existence to achieve. focus identify which of the unit’s products and markets are of interest. conduct audit assess the products and markets identified in focus stage. summarise summarise the products in the business unit in a form suitable as a starting point for the setting of objectives. set objectives set objectives for the business unit based on the information collected, analysed and summarised. set strategy define strategy by which the objectives are to be met. f u r t h v r r e f i n e m e n t o f scope various related areas are outside e x m a r ’s scope, on the grounds that, though important, they are peripheral to the central concerns of exmar, and should not be studied in detail in the interests of timely focus. these areas are summarised in the boxes on the diagram below outside the “scoping” dotted line. brief notes on these follow. r ------------------------- i i i i i i i i i i i i p r o d u c e s t r a t e g i c i m a r k e t i n g p l a n f o r a f b u s i n e s s unit i i i i i s e l e c t / d e f i n e b u s i n e s s unit i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i i ----i d e f i n e unit m i s s i o n [piif=] c o n d u c t auoit s u m m a r i s e s e t o b j e c t i v e s s e t s t r a t e g y ---------we------ i i i i i i i i i i i i i i i i i i i i i i i i w - w - c o r p o r a t e o b j e c t i v e s e n l n g organis- ation d i a g n o s i s m a r k e t r e s e a r c h m a r k e t s e g m e n t a t i o n i p r o o u c t positioning l i organisation diagnosis such issues as diagnosis of the health of an organisation, blake- mouton matrix etc. corporate objective setting the means by which corporate objectives are arrived at is not within exmar's scope. where corporate objectives (or business unit objectives derived from them) are required by later parts of the marketing process, they are regarded as an input to the model. market research, market segmentation, product positioning such techniques as research into the needs or wants of customers, positioning products within markets by finding criteria with which to map the market, and related market segmentation techniques are not covered. the results of, market segmentation are important to the functions modelled, so this is essentially an input to the model, though some assistance may be offered. techniquesconsidered porter matrix critical success factors table directional policy matrix . ansoff matrix boston matrix product life cycle gap analysis objectives typology threat assessment market attractiveness table cost experience curve porter -force model downside risk assessment t e c h n i q u e s left o u t opportunity matrix product positioning m a p customer preference m a p market segmentation m a p diffusion of innovation blake/mouton matrix organisation diagnosis mcdonald productivity matrix size/diversity graph (part of organisation diagnosis) market segmentation studies - detail to investigate financial s u m m a r y - both part of marketing audit response elasticities second level functional breakdown pefine unit mission business unit deffnftfon this involves definition of what the unit is for, including any financial targets. this will be a corporate mission statement if the whole organisation is being considered. otherwise it will identify the specific role of the unit within the organisation. select/define business unit r * financlal define unit . - summary mission wsiness defini- * tion / unit mission this involves defining which business unit the plan is for. where a plan is being produced for an organisational unit, this simply involves identifying the unit. but it may be .more complex: one may wish to carry out the plan just for a subset of an organisational unit’s business of particular interest, or for an area of the business that crosses organisational boundaries. for example, a plan for tinned foods within a foods company may cross department boundaries of design, production, finance, etc. - - i the output is a definition of the business unit, including a title that can be used to head all documents associated with the plan. it m a y be possible to produce a checklist to assist in this function. financial summary any financial targets set for the unit, particularly for revenue or profit. this also involves specification of the planning period to the end of which the targets relate (typically years). business definition/unit mission statement a statement in words to cover aspects of the mission not covered by the financial summary. brief statements should be made which cover the followings points: role or contribution of the unit e . - profit generator, - service department, - opportunity seeker ii) definition of the business - the needs satisfied or the benefits provided. should not be too specific (eg. “we sell milking machinery”) or too general (eg. “we’re in the engineering business”). iii) distinctive competence - this should be a brief statement that applies only to the specific unit. a statement that could equally apply to any competitor is unsatisfactory. . iv) indications for future direction - a brief statement of the principal things that serious consideration would be given to (eg. move into a new segment). focus c o s t experience c u r v e j focus the object is to identify which market segments and products are to be considered in production of the marketing plan. this involves ignoring some detail for the sake of aiding understanding about the critical issues involved. for example, an audit of tinned foods may decide to focus on baked beans and pet foods, and ignore the small market for anchovies. identify % critical to business the basic rule of thumb is that the % of the organisation’s markets and products most critical to its success are those that should be included in a strategic marketing plan. this is a guideline only: the planner may wish to conduct a more or less exhaustive plan. the porter matrix may assist by showing the relative strength of the products in their markets in terms of differentiation and cost leadership, as an indication of the possible future importance of the products. segment the market the relevant markets should be identified and, where appropriate, segmented. this is in general a creative and important step. limited guidance’ only is incorporated in this model. the porter matrix may be of assistance in market segmentation, as clusters of products in similar positions might reasonably be placed in a segment. the standard industrial classification (sic) used as a basis of statistics collection by the government can form a useful starting point for market definition, as a checklist from which to select, though it is not always appropriate. predict next years prediction of the future prospects of the products, all other things being equal, is important as an input into the audit of the current position. it is also an important validation step, as it may affect which % of the products and markets are d,eemed to be critical. for example, if the demand for anchovies is expected to rise steeply in the next three years, it may be decided to include them in the tinned foods audit after all. consideration of where the product is in its life cycle may assist in prediction. the ansoff matrix may already at this point suggest new markets and products that should be defined and considered. the cost experience curve may suggest what is likely to happen to the costs of the products, which may have implications for its future prospects. conduct audit ' c o n d u c t -i audit .s f a c t o r s .s * a s s e s s c r itical - s t r e n g t h s . s u c c e s s a n d f a c t o r s table w e a k n e s s e s i .- o t .s * .v checklist e. o p p o r t u n i t i e s t h r e a t s s_ .- .- ness f a c t o r s .b r- a s s e s s m a r k e t a t t r a c t l v e n e s s the objective is to assess the state and prospects of the products and markets already identified. information needed at this point m a y have been collected in advance of the planning process, or it m a y be collected now. assess strengths and weaknesses the strengths and weaknesses of the company’s products in its markets can be summarised in a critical success factors table. it is very important to get this right, and to validate it against information on the competitors in the market and their strength in the markets. if the information is not available to sufficient accuracy, it should be obtained. after all, one is identif\ .‘irg factors critical to the success of the business. a checklist is available of possible factors to consider. assess opportunities and threats the porter -force model of pressures on you can assist in identification of threats. the threat assessment matrix gives guidance on whether to include ihe threats in the summary list. a checklist of possible opportunities and threats is available. assess market attractiveness the market attractiveness table summarises the attractiveness of a market to the company. it thus complements the critical success factors (csf) table: csf summarises the company’s prospects of success in the market if it chooses to compete, whereas this table summarises the desirability of competing. one important aspect of the market’s attractiveness is the expected future of the market: in this way, the market attractiveness table may be more forward looking than the csf table. - - _- s u m m a r i s e ._____________ i;-ll \ \ \ \ \ \ \ “.‘\f \ -* \ \ x directional \ \ \ x policy matrix \ \ i * m analysis - s u m m a r y i i the objective is to summarise the products in the business unit in a form suitable as a starting point for setting of objectives. the essential component of this is the directional policy matrix, with the current picture of the portfolio, and current projections. the projections can then be modified during the setting of objectives. i the axes of the d p m have already been determined during the audit, being the c s f factors and weightings, and the market attractiveness factors and weightings. guidelines for the reduction of the number of products to be displayed to a sensible number m a y be used; and the axes m a y be changed and/or relabelled in order more effectively to differentiate between products, if initially they are excessively clustered. groups of products, including portfolios, m a y meaningfully be plotted on the d p m , as well as single products: mcdonald gives an example of cranfield school of management’s courses. - c - the boston matrix m a y be used if it is appropriate in this case, on the grounds of its greater simplicity. similar remarks apply to those above about the d p m . a financial gap m a y be ascertained at this point between a unit financial objective and the current projections. this gap is notated on the diagram as a thermometer, as in essence it simply records a gap between two values, though the traditional graphic representation has the advantage of recording the value of a third dimension of the current position. similarly, a “strategic gap” m a y be identified between other objectives of the unit included in the mission statement, and their anticipated fulfilment on the basis of current predictions of the unit’s products and consequent work. this m a y be to do with maintaining the synergy of the organisation. such a strategic gap would be recorded in an analysis summary. s e t o b j e c t i v e s --____---______ the purpose of this function is to produce a list of objectives. these should be quantified, but beyond this the possible types of objectives have not been identified. gap analysis m a y be used to drive this process, by attempting to close the gap starting with productivity improvements, then considering n e w markets ‘and n e w products in the order suggested by the ansoff matrix, and finally by considering changing the business’s assets (changing the nature of the business) in order to meet the objective. at any point, changing the objective m a y also be an option. this is the reason for the feedback line to “define unit mission”. the d p m suggests “directional policy guidelines” for each product/product group plotted on it. these are taken into account in setting objectives for the product or product group. the porter matrix m a y provide further help in this. boston m a y be used, if, again, its implications are acceptable in this case. m/ z ;trategy - * the strategic steps needed to meet the objectives are identified and recorded. the model has not yet been extended to cover this in any more detail. data model notation the data model diagrams presented later in the section are in a format known as entity relationships diagrams. the diagram below is used to explain the notation, is sold into product boxes represent “entities” and lines represent “relationships”. an entity is anything you wish to hold information about, such as products and markets. the information can be represented by blobs by the box, with text describing the information. each item is called an attribute, such as a market’s size. a star in place of a blob indicates an attribute that can be used to identify the particular entity concerned. a relationship represents some connection between the entities. for example, products are related to markets in that a product may be sold into a given market. an arrow leading from entity a to entity b indicates that a given instance of entity a may be related to more than one of entity b. text by the line may be used to indicate the nature of the relationship. so a product may be sold into more than one market, and a market m a y have more than one product sold into it. the case, where there is an arrow at each end, is called a many- to-many relationship. data model ( ) l oeflnttlon p ilfj? l planning / z?zt~on to u n ’t (self or organrrarion cornpet, tor~sf’u~u’e i m a r k e t e g m e n t : i:oz.vtk into (l c) is sold in to p r o j n c t l name . costs newlexistlng (self of competrtor) this gives a simplified data model, as a step towards the full model, described in the next section. products are in a many-to-many relationship with markets they are in, as are business units. products, . markets a n d business units m a y all b e nested within others. composite products are products consisting of several other products, which are sold individually as well. a n example might be a variety pack of cat food. a portfolio is a set of products that, by contrast, is not sold as a set, but which is in s o m e way related. the total range of cat foods offered in an example. if a product is neither of these, it is called a basic product. the critical success factors and weightings that apply to a given market are c o m m o n to all competitors in the market, so they are an attribute of the market itself in the model. this diagram is inadequate when you consider information such as market share. market share is not an attribute of products: a given product m a y be sold into two markets, in each of which it has a different market share. s o a n e w entity is needed ‘between” product and market. similarly, the attractiveness of a market to a given firm is specific to that firm, so a n e w entity is needed between market and business unit. data model (il) *market arramveness / factors ( m a w by l name involve- w a r k e t o r l need fulfilled (defn ) l score on mafs. iz+fi$;;~n ‘-- ‘m a c e ’ y + m e n t i n and overall score b u s ! n e s s / m a r k e t l sac m a r k e t - unit s e g n ient growtn info lc) (actual of p o tential) (self or l oeflnmon i i l c.s.f ‘i l c.s.f ‘i (actual of p o tentw com p e titor) l plannmg l differentlatlon (l- ) ‘r o d u c t f o r l market share . costs score on csfs. t overall score new/exisung l price/average prrce sales volume is sold p r o d u c t * name costs newlexistlng period relation co orgamsation structure i objective (self or compecrcor) a n important area in which this model needs extension is in modelling of features that change over time. this is only loosely described at present, for example by the attribute “growth info” for markets. o n e possibility is to have a different entity for each year (or other period) under consideration - so you might have six market entities, one for each year from three years ago to the end of the planning period in three years’ time, with differing information as to market size and critical success factors. a n intermediate possibility would be to have s o m e information that is static over time, and other information in a separate dynamic entity. this needs investigation. techniaue interrelationshibs patausedby tecmiquu the diagram below shows the data used as input by some of the techniques modelled. ~~~~~~~~~~~~~~ c.s.f ‘s \ ~~~:~~ \ *. \ (actual or pote$al) competitor) l phlnm g ! l aclarion tc l differentiatlqh (l- ) . . : i l market share’ *. i product for . costs ---_ l score on csf;,-~‘oweralj&re l new/existing ., t pnce/average prq sales volume -. *. i ‘\ i ’ , -. i is sold in to .’ product . ‘\ ** +q / ’ : ii i i i i i i i’ : , i , i i i i i i i i i i i i i , , i i i i l name: l name: : : i i costs’ : costs’ : new/existind new/existind (self or compethor) techniaue interrelationshins the diagrams below show various connections identified between techniques. they assume that by using a technique, any data required by it is entered into the model by s o m e means, so that data is available for another technique. t e c h n i q u e interrelationships ( ) horuontal a m vemcal a m poky guldelmes current growh experience + : provides x as input to - technique interrelationships (ii) dwemonal pohcy gwdehner c\ i xl b o s t o n qap i n i d~fferwm~twn. to atd jo forecastmq c o s t experience c u r v e - - pii: - ( ) - o(w - $ . + .oo iyx pergamon preu ltd csrl: a language for expert systems for diagnosist tom bylander, sanjay mittal* and b. chandrasekaran artificial intelligence group, department of computer and information science, the ohio state university, columbus, oh , u.s.a. (received september ) abstract-we present csrl (conceptual structures representation language) as a language to facilitate the development of expert diagnosis systems based on a paradigm of “cooperating diagnostic specialists.” in our approach diagnostic reasoning is one of several generic tasks, each of which calls for a particular organizational and problem-solving structure. a diagnostic structure is composed of a collection of specialists, each of which corresponds to a potential hypothesis about the current case. they are organized as a classification or diagnostic hierarchy, e.g. a classification of diseases. a top-down strategy called establish-refine is used, in which either a specialist establishes and then refines itself, or the specialist rejects itself, pruning the hierarchy that it heads. csrl is a language for representing the specialists of a diagnostic hierarchy and the diagnostic knowledge within them. the diagnostic knowledge is encoded at various levels of abstractions: message procedures, which describe the specialist’s behavior in response to messages from other specialists; knowledge groups, which determine how data relate to features of the hypothesis; and rule-like knowledge, which is contained within knowledge groups. i. introduction many kinds of problem-solving for expert systems have been proposed within the ai community. whatever the approach, there is a need to acquire the knowledge in a given domain and implement it in the spirit of the problem-solving paradigm. reducing the time to implement a system usually involves the creation of a high-level language that reflects the intended method of problem-solving. for example, emycin[ was created for building systems based on mycin- like problem-solving[ ]. such languages are also intended to speed up the knowledge acquisition process by allowing domain experts to input knowledge in a form close to their conceptual level. another goal is to make it easier to enforce consistency between the expert’s knowledge and its implementation. csrl (conceptual structures representation language) is a language for implementing expert diagnostic systems that are based on our approach to diagnostic problem-solving. this approach is an outgrowth of our group’s experience with mdx, a medical diagnostic program[ ], and with applying mdx-like problem-solving to other medical and nonmedical domains. csrl facilitates the development of diagnostic systems by supporting constructs that represent di- agnostic knowledge at appropriate levels of abstraction. first, we will overview the relationship of csrl to our overall theory of problem-solving types and the diagnostic problem-solving that underlies csrl. we then present csrl, illus- trating how its constructs are used to encode diagnostic knowledge. two expert systems under development in our laboratory, which use csrl, are then briefly described. based on our experience with these systems, we point out where improvements in csrl are needed. . classificatory diagnosis the central problem-solving of diagnosis, in our view, is classificatory activity. this is a specific type of problem-solving in our approach, m;aning that a special kind of organization and special strategies are strongly associated with performing expert diagnosis. in this section we will briefly review the theory of problem-solving types, as presented by chandrasekaran[ ], and the structure and strategies of the diagnostic task[ ]. tthis an expanded version of a paper of the same title presented at the international joint conference on artificial intelligence. wurrently at knowledge systems area, xerox parc, coyote hill rd., palo alto, ca , u.s.a. . types of problem-solving t. bylander et al. we propose that expert problem-solving is composed of a collection of different problem- solving abilities. the ai group at ohio state has been working at identifying well-defined types of problem-solving (called generic tasks), one of which is classificatory diagnosis. (for the purposes of this discussion we will use “diagnosis ” in place of “classificatory diagnosis” with the understanding that the complete diagnostic process includes other elements as well.) other examples include knowledge-directed data retrieval, consequence finding, and a restricted form of design. each generic task calls for a particular organizational and problem-solving structure. given a specific kind of task to perform, the idea is that specific ways to organize and use knowledge are ideally suited for that task. even when the specification of a problem is reduced to a given task within a given domain, the amount of knowledge that is needed can still be enormous (e.g. diagnosis in medicine). in our approach the knowledge structure for a given task and domain is composed of specialists, each of which specialize in different concepts of the domain. domain knowledge is distributed across the specialists, dividing the problem into more manageable parts, and organizing the knowledge into chunks that become relevant when the corresponding concepts become relevant during the problem-solving. decomposing a domain into specialists raises the problem of how they will coordinate during the problem-solving process. first, the specialists as a whole are organized primarily around the “subspecialist-of” relationship. each task may specify additional relationships that may hold between specialists. second, each task is associated with a set of strategies that take advantage of these relationships and the problem-solving capabilities of the individual specialists. the choice of what strategy to follow is not a global decision, but is chosen by the specialists during problem-solving. . the diagnostic task the diagnostic task is the identification of a case description with a specific node in a predetermined diagnostic hierarchy. each node in the hierarchy corresponds to a hypothesis about the current case. nodes higher in the hierarchy represent more general hypotheses, and lower nodes are more specific. typically, a diagnostic hierarchy is a classification of malfunc- tions of some object, and the case description contains the manifestations and background information about the object. for example, the auto-mech expert system[ ] attempts to classify data concerning an automobile into a diagnostic hierarchy of fuel-system malfunctions. figure illustrates a fragment of auto-mech’s hierarchy. the most general node, the fuel system in this example, is the head node of hierarchy. more specific fuel-system malfunctions, such as fuel-delivery problems, are classified within the hierarchy. each node in the hierarchy is associated with a specialist that contains the diagnostic knowledge to evaluate the plausibility of the hypothesis from the case description. from this knowledge the specialist determines a confidence value representing the amount of belief in the hypothesis. if this value is high enough, the specialist is said to be established. the basic strategy of the diagnostic task is a process of hypothesis refinement, which we call establish-refine. in this strategy, if a specialist establishes itself, then it refines the hypothesis by invoking its subspecialists, which also perform the establish-refine strategy. if its confidence value is low, the specialist rejects the hypothesis and performs no further actions. note that when this happens, the whole hierarchy below the specialist is eliminated from consideration. otherwise the specialist suspends itself and may later refine itself if its superior requests it. bad fuel problems n--._ fuel mixture problems low octane water in fuel dirt in fuel fig. i fragment of a diagnostic hierarchy csrl with regard to fig. , the following scenario might occur. first, the fuel-system specialist is invoked, since it is the top specialist in the hierarchy. this specialist is then established, and the two specialists below it are invoked. bad fuel problems are rejected, eliminating the three subspecialists of bad fuel from consideration. finally, the fuel-mixture specialist is established, and its subspecialists (not shown) are invoked. an important companion to the diagnostic hierarchy is an intelligent data-base assistant that organizes the case description, answers queries from the diagnostic specialists, and makes simple inferences from the data[ ]. for example, the data base should be able to infer that the fuel tank is not empty if the car can be started. the diagnostic specialists are then relieved from knowing all the ways that a particular datum could be inferred from other data. there are several issues relevant to diagnostic problem-solving that we will not address here. the simple description above does not employ strategies for bypassing the hierarchical structure for common malfunctions, for handling multiple interacting hypothesis, or for ac- counting for the manifestations. also, additional control strategies are required when many nodes are in a suspended state. for discussion on some of these topics, see gomez and chan- drasekaran[ ]. test ordering, causal explanation of findings, and therapeutic action do not directly fall within the auspices of the classificatory diagnosis as defined here, but expertise in any of these areas would certainly enhance a diagnostic system. fully resolving all of these issues and integrating their solutions into the diagnostic framework are problems for future research. . differences from other approaches the usual approach to building knowledge-based systems is to emphasize a general knowl- edge representation structure and different problem-solvers that use that knowledge. one dif- ference in this approach is that the organization of knowledge is not intended as a general representation for all problems. rather it is tuned specifically for diagnosis. by limiting the type of problem to be solved, a specific organizational technique (classification hierarchy) and problem-solving strategy (establish-refine) can be used to provide focus and control in the problem-solving process. another difference is that the specialists in the hierarchy are not a static collection of knowledge. the knowledge of how to establish or reject is embedded within the specialists. each specialist can then be viewed as an individual problem-solver with its own knowledge base. the entire collection of specialists engages in distributed problem-solving. . csrl csrl is a language for representing the specialists of a diagnostic hierarchy and the diagnostic knowledge within them. the diagnostic knowledge is encoded at various levels of abstractions. message procedures describe the specialist’s behavior in response to messages from other specialists. these contain the knowledge about how to establish or refine a specialist. knowledge groups determine how selected data relate to various features or intermediate hy- potheses that are related to the specialist. the selected data may be the values of other knowledge groups, so that a single knowledge group can “summarize” the results of several others. knowledge groups are composed of rule-like knowledge that matches the data against specific patterns and, when successful, provides values to be processed by the knowledge group. . specialists in csrl a diagnostic expert system is implemented by individually defining each specialist. the super- and subspecialists of the specialist are declared within the definition. figure is a (specialist badfuel (declare (superspecialist fuelsystem) (subspecialists lowoctane waterinfuel dirtinfuel)) (kgs . ..i (messages . ..)i fig. . skeleton specialist for badfuel t. bylander et al skeleton of a specialist definition for the bad fuel node from fig. . the declare section specifies its relationships to other specialists. the other sections of the specialist are examined below. since csrl is designed to use only a simple classification tree, many choices concerning the composition of the hierarchy must be made. this is a pragmatic decision, rather than a search for the “perfect” classification tree. the main criteria for evaluating a classification is whether enough evidence is normally available to make confident decisions. to decompose a specialist into its subspecialists, the simplest method is to ask the domain expert what subhy- potheses should be considered next. usually the subspecialists will differ from one another based on a single attribute (e.g. location, cause). for further discussion on this and other design decisions in csrl, see bylander and smith[ ]. . messuge procedures the messages section of a specialist contains a list of message procedures that specify how the specialist will respond to different messages from its superspecia ist.t “establish,” “re- fine, ” “establish-refine” (combines establish and refine), and “suggest” are predefined messages in csrl; additional messages may be defined by the user. below, we will examine how establish and refine procedures are typically constructed. message procedures are the highest level of abstraction for diagnostic knowledge within specialists. just as in general message-passing languages, messages provide a way to invoke a particular kind of response without having to know what procedure to invoke. strategies for diagnosis, such as establish-refine, are usually easy to translate into a message protocol. however. csrl does not provide any way to specify and enforce message protocols. figure illustrates the establish message procedure of the badfuel specialist. “relevant” and “summary” are names of knowledge groups of badfuel. “self” is a keyword that refers to the name of the specialist. this procedure first tests the value of the relevant knowledge group. (if this knowledge group has not already been executed, it is automatically executed at this point.) if it is greater than or equal to , then badfuel’s confidence value is set to the value of the summary knowledge group, or else it is set to the value of the relevant knowledge group. in csrl a confidence value scale of - to + is used (integers only). a value of + or + indicates that the specialist is established. in this case the procedure corresponds to the following diagnostic knowledge. first perform a preliminary check to make sure that badfuel is a relevant hypothesis to hold. if it is not (the relevant knowledge group is less than ). then set badfuel’s confidence value to the degree of rclcvancy. otherwise, perform more complicated reasoning (the sum- mary knowledge group combines the values of other knowledge groups) to determine badfuel’s confidence value. figure shows a refine procedure that is a simplified version of the one that badfuel uses. “subspecialists” is a keyword that refers to the subspecialists of the current specialist. the procedure calls each subspecialist with an english message. $ if the subspecialist establishes itself (t ? tests if the confidence value is + or + ), then send it a refine message. csrl has a variety of other kinds of statements and expressions so that more complicated (establish (if (ge relevant ) then (setconfidence self sunrmary) else (setconfidence self relevant))) fig. . establish procedure of badfuel. +a specialist i\ not allowed to send messages to its superspecialist. however. other message-passing routes are allowed. specifically. a specialist may send a message to itself. across the hierarchy, and to indirect subspecialists. in the latter case each interconnecting specialist is sent a “suggest” message and decides within its suggest message procedure whether or not to pass the original message downwards. $‘for convenience many of csrl’s control constructs mimic those of interlisp; however, these constructs are executed by the csrl interpreter, not by using lisp eval. lisp code is allowed within message procedures, but only within a construct called “dolisp. ” this is not intended to let specialists have arbitrary code, but to allow interaction with other lisp-implemented systems. csrl (refine (for specialist in subspecialists do (call specialist with establish) (if (+? specialist) then (call specialist with refine)))) fig. . refine procedure. strategies can be implemented. for example, a “reset” statement deletes the confidence value and the knowledge group values of a specialist. this might be used when additional tests are performed, making it necessary to recalculate the confidence value. also, messages can be parameterized, and message procedures can declare local variables. . knowledge groups the kgs section of a specialist definition contains a list of knowledge groups that are used to evaluate how selected data indicate various features or intermediate hypotheses that relate to specialist’s hypothesis. a knowledge group can be thought of as a cluster of production rules that map the values of a list of expressions (boolean and arithmetic operations on data) to some conclusion on a discrete, symbolic scale. different types of knowledge groups perform this mapping differently: e.g. directly mapping values to conclusions, or ha;ling each rule add or subtract a set number of “confidence” units. knowledge groups are intended for encoding the heuristics that a domain expert uses for inferring features of an hypothesis from the case description. the main problem is that this inference is uncertain--there is rarely a one-to-one mapping from data to the features of the hypothesis. the way that this is handled in csrl is borrowed from the uncertainty handling techniques used in mdx[ ]. each feature or intermediate hypothesis is associated with a knowledge group. the data that the domain expert uses to evaluate the feature are encoded as expressions in the knowledge group. these are usually queries to a separate data-base system. each combination of values of the expressions is then mapped to a level of confidence as determined by the domain expert. this set of knowledge groups becomes the data for another knowledge group, which determines the confidence value of the specialist from the confidence values of the features.? by examining the results of test cases, we see that the knowledge groups are relatively easy to debug, since the attention of the domain expert can be directed to the specific area of knowledge that derived the incorrect result. as an example, fig. is the relevant knowledge group of the badfuel specialist mentioned above. it determines whether the symptoms of the automobile are consistent with bad fuel problems. the expressions query the user (who is the data base for auto-mech) for whether the car is slow to respond, starts hard, has knocking or pinging sounds, or has the problem when accelerating. ‘ ‘askynu?” is a lisp function that asks the user for a y, n, or u (unknown) answer from the user, and translates the answer into t, f, or u, the values of csrl’s three- (relevant table (match (askynu? “is the car slow to respond”) (askynu? “does the car start hard”) (and (askynu? “do you hear knocking or pinging sounds”) (askynu? “does the problem occur while accelerating”)) with (if t ? ? then - elseif ? t ? then - elseif ? ? t then else ))) fig. . relevant knowledge group of badfuel tactually, any number of knowledge group levels can be implemented t. bylander ef al. valued logic. each set of tests in the if-then part of the knowledge group is evaluated until one matches. the value corresponding to this “rule” becomes the value of the knowledge group. for example, the first rule tests whether the first expression is true (the “?” means doesn’t matter). if so, then - becomes the value of the knowledge group. otherwise, other rules are evaluated. the value of the knowledge group will be if no rule matches. this knowledge group encodes the following diagnostic knowledge: if the car is slow to respond or if the car starts hard, then badfuel is not relevant in this case. otherwise, if there are knocking or pinging sounds and if the problem occurs while accelerating, then badfuel is highly relevant. in all other cases badfuel is only mildly relevant. figure is the summary knowledge group of badfuel. its expressions are the values of the relevant and gas knowledge groups (the latter queries the user about the temporal relationship between the onset of the problem and when gas was last bought). in this case, if the value of the relevant knowledge group is and the value of the gas knowledge group is greater than or equal to , then the value of the summary knowledge group (and consequently the confidence value of badfuel) is , indicating that a bad-fuel problem is very likely. . comparison with rule-bused languages there is nothing in csrl that is not programmable within rule-based languages such as opss[lo] or emycln[l]. the difference between csrl and these languages is that csrl makes a commitment to a particular organizational and programming style. csrl is not intended to be a general-purpose representation language, but is built specifically for the classificatory diagnosis problem. it is possible to program in a rule-based language, so that there is an implicit relationship between rules so that they correspond to knowledge groups and specialists. rl, although not a diagnostic expert system, is an excellent example of how one creates implicit grouping of rules in such a system[ . the central idea underlying csrl is to make these relationships explicit. the expert system implementor is then relieved from trying to impose an organization on a organizationless system and is free to concentrate on the conceptual structure of the domain. also, there is a greater potential to embed explanation and debugging facilities that can take advantage of the expert system organization. . the csrl environment the current version of csrl is implemented in interlisp-d and loops, an object- oriented programming tool. each specialist is implemented as a loops class, which is in- stantiated for each case that is run. the loops class hierarchy is used to specify default message procedures and shared knowledge groups, making it easy to encode a default establish- refine strategy, and letting the user incrementally modify this strategy and add strategies as desired. a graphical interface displays the specialist hierarchy and, through the use of a mouse, allows the user to easily access and modify any part of the hierarchy. additional facilities for debugging and explanation are being implemented. . expert systems that use csrl . auto-mech auto-mech is an expert system that diagnoses fuel problems in automobile engines[ ]. this domain was chosen to demonstrate the viability of our approach to nonmedical domains, as well as to gain experience and feedback on csrl.t the purpose of the fuel system is to deliver a mixture of fuel and air to the air cylinders of the engine. it can be divided into major subsystems (fuel delivery, air intake, carburetor, vacuum manifold) that correspond to initial hypotheses about fuel-system faults. tauto-mech was developed using an early version of the language csrl (sunmrary table (match relevant gas with (if (ge ) then elseif (ge ) then elseif ? (lt ) then - ))) fig. . summary knowledge group of badfuel auto-mech consists of csrl specialists in a hierarchy that varies from four to six levels deep. its problem-solving closely follows the establish-refine strategy. before this strategy is invoked, auto-mech collects some initial data from the user. this includes the major symptom that the user notices (such as stalling) and the situation when this occurs (e.g. accelerating and cold engine temperature). any additional questions are asked while auto-mech’s specialists are running. the diagnosis then starts and continues until the user is satisfied that the diagnosis is complete. the user must make this decision because the data that auto-mech uses are very weak at indicating specific problems, and, more importantly, auto-mech is unable to make the repair and determine whether the problem has been fixed. a major part of auto-mech’s development was determining the assumptions that would be made about the design of the automobile engine and the data that the program would be using. different automobile engine designs have a significant effect on the hypotheses that are considered. a carbureted engine, for example, will have a different set of problems than a fuel- injected engine (the former can have a broken carburetor). the data was assumed to come from commonly available resources. the variety of computer analysis information that is available to mechanics today was not considered, in order to simplify building auto-mech. . red red is an expert system whose domain is red-blood-cell antibody identification[l ]. an everyday problem that a blood bank contends with is the selection of units of blood for transfusion during major surgery. the primary difficulty is that antibodies in the patient’s blood may attack the foreign blood, rendering the new blood useless as well as presenting additional danger to the patient. thus, identifying the patient’s antibodies and selecting blood that will not react with them is a critical task for nearly all red-blood transfusions. the red expert system is composed of three major subsystems, one of which is implemented in csrl. the non-csrl subsystems are a data base, which maintains and answers questions about reaction records (reactions of the patient’s blood in selected blood samples under a variety of conditions), and an overview system, which assembles a composite hypothesis of the anti- bodies that would best explain the reaction record[ . csrl is used to implement specialists corresponding to each antibody that red knows about (about of the most common ones) and to each antibody subtype (different ways that the antibody can react). the major function of the specialists is to rule out antibodies and their subtypes whenever possible, thus simplifying the job of the overview subsystem, and to assign confidence values, informing overview of which antibodies appear to be more plausible. the specialists query the data base for information about the test reactions and other patient information, and also tell the data base to perform certain operations on reaction records. an interesting feature of red is the way it handles the problem of interacting hypotheses. it is possible for the patient’s blood to have practically any number or combination of antibodies, which makes it very hard for a single specialist to determine how well it will fit with other specialists in a composite hypothesis. in red each specialist is encoded to assume that it is independent-it looks at the data as if no other specialist can account for the same data. the knowledge of how the specialists can interact is left to the overview subsystem. this would be problematic if few specialists could rule themselves out, but it so happens that in this domain it is rare to have more than a few antibodies that cannot be independently ruled out. thus red’s csrl subsystem makes overview’s problem-solving computationally feasible since it consid- erably reduces the amount of search that would otherwise be necessary. t. bylander et al. . needed improvements in csrl the largest flaw in csrl is that there is no strategy that determines when diagnosis should stop. currently, the default procedures simply ask the user if the current diagnosis is satisfactory. some notion of what it means to account for the data needs to be added to the language. the work on red’s overview system is a step in this direction, but there needs to be more integration of overview and csrl (currently overview starts after the specialists are finished) and a better understanding of what kinds of interactions can occur between two hypotheses. progress in this area would also help increase the focus of the diagnosis; i.e. the diagnosis could concentrate on accounting for the most important manifestation(s). another problem is the meaning of the confidence value of a specialist. in mdx this value was directly associated with the amount of belief in the specialist. however, in both auto- mech and red, this meaning had to be slightly altered to fit the purposes of the expert system. in auto-mech the confidence value is used to indicate whether the hypothesis was worth pursuing. in red it is used to indicate the specialist’s plausibility, given the independence assumption mentioned earlier. it is not possible in either expert system to confirm a specialist without outside help. in auto-mech a repair or highly specific test must be performed, but in red all the specialists must be considered together. this does not create a problem for the process of establish-refine problem-solving, but makes it difficult to explain what the confidence value means. any explanation facility must understand the assumptions that are being made in order to make coherent explanations. . conclusion we believe that the development of complex expert systems will depend on the availability of special-purpose languages with organizational and problem-solving tools that match the conceptual structure of the domain. csrl represents an initial step in this direction. it provides facilities to organize diagnostic knowledge in accordance with the structure of the domain. in particular, csrl’s constructs facilitate the encoding of rule-like and strategic knowledge into appropriate abstractions: knowledge groups, message procedures, and specialists. acknow~/edgments~we would like to acknowledge jack smith and jon sticklen for many fruitful discussions concerning csrl’s design. many improvements in the language are due to mike tanner and john josephson, who implemented the csrl specialists in auto-mech and red., the language development is funded by a grant from the battelle memorial laboratories university distribution program, and experimentation and application in different domains is supported by afosr grant - , and nsf grant mcs- . i . io. ii. . . references w. van melle, a domain independent production-rule system for consultation programs. proc. sixth int. cot$ on artijzcial intelligence, pp. - . tokyo ( ). e. h. shortliffe, computer-based medical consultations: mycin. elsevier. new york ( ). b. chandrasekaran and s. mittal, conceptual representation of medical knowledge for diagnosis by computer: mdx and related systems, in advancrs in computers, pp. - . academic press. new york, ( ). b. chandrasekaran, towards a taxonomy of problem solvin, ~ types. ai mq. . ( ). f. gomez and b. chandrasekaran, knowledge org,;nization and distribution for medical diagnosis. ieee trans. svst.. man cybentetics smc- . - ( ). m. c. tanner and ‘f. bylander, application of the csrl language to the design of expert diagnosis systems: the auto-mech experience. proc. joint swvices workshop on artificial /nte//igence in maintenunce. department of defense, pp. lll , denver ( ). s. mittal and b. chandrasekaran, conceptual representation of patient data bases. j. medical swt. , - ( ). t. bylander and j. w. smith, using csrl for medical diagnosis. pmt. .%wmd int. cmfl ivi medical computer science and compututimal medicine. ieee computer sot.. glouster. ohio c ). b. chandrasekaran, s. mittal and j. w. smith, reasoning with uncertain knowledge: the mdx approach. proc. con,qrcts americtm medical informtics a.ssociation. san francisco ( x ). c. l. forpy, ops users manuul. tech. rept. cmu-cs- i , carnegie-mellon university ( i i). j. mcdermott. ri: a rule-based configurer of computer systems. artificial irttrll. iy. - ( ). j. w. smith. j. josephson, c. evans, p. straum and j. noga, design for a red-cell antibody identification expert. proc. srcond lat. co@ on medical computer science and computational mrdir,inr. ieee computer sot., glouster, ohio ( ). j. josephson, b. chandrasekaran and j. w. smith, the ovrrvte~~ fmctim in diagnostic problem solving. technical paper, al group, dept. of computer and information science, the ohio state university ( ). by a. terry bahill, pat n. harris, and erich senn sometimes what you learn building an expert system is more important than whether or not it succeeds lessons learned t^% * * t—i building expert systems w e have made many expert sys- tems. two of our best are co- gito, which gives installation advice for bringing up the u n i x (bell laboratories) . bsd operat- ing system on a vax (digitial equipment corp.), and data communication diagnosis (dcd), which helps a person connect a ter- minal to a computer. this article w i l l com- pare and contrast the development, tool se- lection, and testing of these two systems. background before cogito was completed, two detailed reference manuals were used for u n i x . bsd i n s t a l l a t i o n instructions. these vo- luminous manuals were difficult to use. in contrast, cogito is easy to use because it fil- ters the information and presents only rel- evant advice about the user's computer sys- tem. cogito remembers data the user has previously entered and uses it to customize its response. cogito is w r i t t e n in m . i (tek- nowledge inc.), runs on a personal com- puter, and uses if-then production rules to encode the task knowledge. dcd is a knowledge-based system de- signed to help engineers connect a periph- eral device (terminal or printer) to a com- puter, modem, or local area network. most of dcd was w r i t t e n in ops , although parts were written in c and m . i . instructions for connecting peripheral de- vices to computers are contained in the ref- erence and user manuals of the devices. un- fortunately, their usage is fraught with difficulty. it is not easy to collect all the rel- evant information for the two devices in question since this information is spread over several chapters (for example, pin con- nections are discussed in the hardware sec- tion and transfer parameters in the set-up section) in several manuals written by differ- ent companies. dcd had three specific requirements: • provide enough advice to people who have some previous computing experience but little experience in data communica- ai expert • september tions so they can design a cable with a - pin connector on each end that can be used to connect a peripheral device to a host device • read in a description (name) of the de- vices and ask the user for values of param- eters not stored in the expert system's data base, provide a help feature to support the process of answering questions, and produce a sequence of instructions to help the user successfully interconnect the devices • derive an approximate solution even if some information is missing or the answers to questions are unknown. unfortunately, modern computer users are often overwhelmed with information. expert systems can help reverse this trend. voluminous data with many interrelation- ships presented over a short time span can overload the user. data become useful in- formation when the relationships between items are cohesive. both cogito and dcd eliminate extraneous information and relate relevant information, reducing the chances of user overload. creating the knowledge base our first expert system, made in , had a flat tree: every premise was tied directly to the conclusion. however, to emulate human input-output behavior, we believed that an expert system should have a bushy tree. we discovered that we had to invent interme- diary conclusions if we were to mimic a con- sultation with a human. furthermore, hu- mans have a limited short-term memory, thus a solution going from premises to con- clusions should follow a path of simple knowledge intermediates. listing shows the hierarchy we used to make a bushy tree for dcd. of course, even with this hierarchy, the rules do not wrrite themselves; the expert still had to add knowledge to get the rules. for example, the expert knew that pins and are used to transmit and receive data. so if pin transmits data, pin receives data. the if-then production rules shown in fig- ure were derived from the hierarchy of listing . the hierarchy shown in listing helped make sure rules were not over- looked. for example, it was easy to check that each pin was contained in at least one rule. this hierarchy was not very structured, but it still helped in writing the rules. rules are easier to write in problem domains that can be structured more systematically with a technique that divides the problem space into a set of objects where every object has attributes and every attribute has values. this object-attribute-value ordered set is called an o-a-v triplet. typical usage of o-a-v triplets is: the ob- ject has an attribute and a value is an a t t r i - bute. figure shows a knowledge base re- presented in the o-a-v schema. the lower left triplet in this knowledge base can be written as "coat has markings" and "striped is a marking." another way of writing the o-a-v triplet is as a sentence ("attribute of object is val- ue"); for example, "marking of coat is striped." this usage makes the construction of facts easy: ai expert • september description-parameters name [a..z], [a..z], [-] sytek-unit-number xxxx.x --> [ .. ] general-parameters baud rate to , unknown standard rs- c, rs- , centronics, unknown synchronization asynchronous, synchronous, unknown duplexity full-duplex, half-duplex, simplex, unknown hardware-protocol not-required, required, unknown software-protocol none, xon/xoff, ack/nak, unknown echo on, off, unknown function mapping - . - . unknown pinl frame-ground, not-available, unknown pin transmit-data, receive-data. unknown pin receive-data. transmit-data, unknown pin request-to-send, not-available, unknown pin. clear-to-send, not-available, unknown pin data-set-ready, not-available, unknown pin signal-ground, not-available, unknown pin data-carrier-detect, not-available, unknown pin data-terminal-ready, not-available, unknown character-format data-bits , . unknown stop-bits , . unknown parity none, even. odd. unknown listing . parameter hierarchy. figure . if-then production rules derived from the hierarchy in listing . type of animal is mammal category of mammal is ungulate extremities of ungulate is hooves. w i t h such facts, it is easy to write if-then pro- duction rules, for example: if composition of coat is hair goal = advice. question(computer-pin ) = 'what is the function of pin on the computer side?'. iegalvals(computer-pin ) = [transmit-data. receive-data. unknown]. if computer-pin = transmit-data then computer-pin = receive-data. if computer-pin = receive-data and terminal-pin = receive-data then sketch = diagraml. if sketch = diagraml and display(['interconnect the data signals like this ± '.nl .nl]) then advice then type of animal is mammal. if type of animal is mammal and extremities of animal is hooves then category of animal is ungulate. if category of animal is ungulate and marking of coat is stripes then identity of animal is zebra. a tree is seldom built all at one time; in- stead, branches are added incrementally. when a branch is added, sometimes some- thing that is a value at one level becomes an object at a lower level ("mammal" is a value in "type of animal is mammal," but becomes an object in "category of mammal is ungu- late.") this o-a-v = o-a-v linkage is shown in the right column of figure . in contrast, sometimes o-a-v triplets are linked together without an intervening attribute: type of animal is mammal marking of coat is stripes. this o-a-v-o-a-v linkage is shown in the left column of figure . teknowledge distributes a tutorial expert system, sacon, with their pc-based shell, m . i . the company has been refining sa- con for years. it is composed exclusive- ly of o-a-v = o-a-v linkages like those in the right column of figure . this implies that perhaps these are the best linkages for an expert system. we have subsequently tried several other methods for constructing if-then production rules without forcing human experts to dis- tort their reasoning processes to produce if- then-type rules. we have found the analytic hierarchy process -' helpful in constructing o-a-v triplets. selecting the proper tool consider the following analogy. a sculptor wishes to create an image of a dove, and has both marble and ice in which to carve. al- though the mediums of ice and marble re- quire different tools and speed of construc- tion, the marble and ice doves are the same as the image within the sculptor's mind. the difference in mediums does not affect the result. now suppose the sculptor wants to sculpt with iron rebar and a torch. this dove will not appear the same. in this case, the medium does affect the result. current expert system shells force a struc- ture on the knowledge base that may be dif- ferent from the expert's structure; for ex- ample, requiring the knowledge to appear as production rules, semantic networks, or frames. when an expert has trouble devel- oping a rule that covers a certain situation, ai expert • september it is often said that the expert's knowledge is nonverbal and internalized. alternatively, perhaps the expert cannot develop a rule that fits the situation because he or she does not t h i n k in terms of if-then production rules. cogito provided m a n y examples of how humans are forced to fit their knowledge into the constraints of a tool. cogito has many rules that m i g h t otherwise be simple lists. furthermore, cogito's problem do- main is probably most suitable for a for- ward-chaining strategy. if the inference en- gine had been capable of extensive forward- chaining, the knowledge base could have been shortened. thus, the tool used defi- nitely affected the coding of the knowledge. however, an engineer's job is to find and apply the best available tool to solve a prob- lem. the tools available for cogito were m . i and ops . the decision to use m . i was primarily based on the differences in rule implementations and ease of i n p u t and out- put. cogito's rules are english-like phrases. ops rules are reminiscent of lisp code; experts who are not f a m i l i a r with ops or lisp have trouble reading and understand- ing ops rules. rule readability is impor- tant so the knowledge engineer can verify with the expert that the intent of the rule is the same as the coded rule. with m . i , the knowledge engineer can concentrate on problem semantics and not on the syntax of ops or lisp. we selected ops for dcd primarily be- cause it used forward chaining, and this seemed the best type of inferencing when all the data are available at the beginning of the inferencing process. however, a subset of the problem (creating a cable plan) was also written in m . i , which gave us the op- portunity to compare these two tools. the q u a l i t y of a system's user interface is crucial for a system's survival. m . i has an advantage in this matter since it provides le- gal value checking, which is d i f f i c u l t and cumbersome to implement in ops . the human-computer interface of m . i is better than t h a t of ops , but this did not matter in our case since all the data were collected from the human expert before the infer- ence process started using a module written in c. the knowledge base contains the rules to solve the problem at hand. rules written in m . i are easier to read than those in ops . however, the easier rule syntax of m . i is only an advantage for novice knowledge en- gineers. based on our experiences, we think experienced programmers can write either m . i or ops rules with equal ease. on the other hand, m . i is at a disadvantage be- cause its conclusions cannot have values for several different expressions. this handicap is not shared by ops , where several differ- ent expressions can be in the conclusion of a animal type composition carnivoremarking striped spotted ungulate extremities hooves v- v-o figure . object-attribute- value hierarchy. ai expert • september table . criteria and attributes used by the multi-attribute utility technique to help select an expert system shell; w— weight, r—rating, ra—ranking. rule, reducing the number of rules. m . i automatically handles the answer "unknown." in ops , we had to specify in each rule how "unknown" should be han- dled. the inference engine of m . i uses cer- tainty factors. ops , on the other hand, does not provide a mechanism that deals w i t h uncertainty. probability calculations can be done in ops , but this is cumber- some. however, dcd, like most expert sys- tems, did not need certainty factors. in fact, in one of our expert system classes at the university of arizona, tucson, we found that only six of the student-generated systems used certainty factors. quantitative selection we have helped build over expert sys- tems for class projects, master's theses, and commercial ventures. we are often asked why we chose a particular expert system shell. often our qualitative answers do not satisfy the questioner. therefore, for the dcd project, we provided a quantitative an- swer to this question. no one-to-one match exists between problems and software tools. however, con- ceptions of particular types of problems are the same for several different domains. these conceptions are commonly referred to as consultation paradigms. examples of consultation paradigms are diagnosis/ad- vice, planning, and design. diagnosis/ad- vice is the consultation paradigm used by dcd. according to p. harmon and d. king, the shells the best suited for diagnosis/ad- vice are m . i , s . i , and ops . hundreds of commercially available expert system shells criteria and attributes w r ra representation facts . if-then rules . inference strategy . uncertainty . implementation software available . hardware available . data base access . interfacing user display . explanation facilities . help functions . editor . debugging aid . support documentation . courses . applications . summation — — m. s. ops . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . exist; we did not consider every one for dcd, but restricted ourselves to these three. the following discussion illustrates a technique and is not meant to advocate a particular shell. to select the shell for our data-communi- cation problem, we concentrated on the fol- lowing criteria (table ): • representation: the knowledge that en- ables us to solve data-communication prob- lems must be formulated with if-then pro- duction rules. the inference strategy used should be forward-chaining since all neces- sary data are given at the beginning of the inference process. the tool to be selected must provide an approximate solution when some information is missing or the answers to questions are unknown. • implementation: when this project be- gan, available hardware included several personal computers, many pdf l l s , and a vax. available software included ops , m . i , a n d rosie. • interfacing: questions asked by the sys- tem should be displayed on a screen and an- swers entered on a keyboard. solutions to a problem must be written to a data file. the knowledge base must be built and modified with a tool-independent text editor. • support: documentation should be rea- dable with easy-to-understand examples. re- ferences to past tool applications and their level of success should be available. many software packages that are com- mercially available (such as expert choice ) aid multiobjective decision analysis in the face of nebulous and uncertain data. we used the multi-attribute u t i l i t y technique (maut), sometimes called smart, to choose among m . i , s . i , and ops . the re- sults are shown in table . in 'table , the importance of a particular description criterion is expressed with a rank (ra) from to , which is then scaled into the rating (r) with a range of to . for example, rank was assigned to the cri- terion implementation since the required software and hardware had to be available to finish the project on time. the weight (w) of an attribute is its indi- vidual rating divided by the sum of all the ratings. when the weights have been de- rived for the attributes, it is time to assign values (between and ) for the attributes to the three software systems being com- pared. for example, the attribute software available has an assigned value of . for ops since all the necessary software was up and running. s.i has a value of . be- cause we did not have it and could not af- ford to buy it. next the weights are multiplied by the as- signed values and the sums placed at the bottom of the columns for m . i , s . i , and ops . the best tool for the job was the one ai e x p e r t • september . . . initializing /etc/fstab change into the directory /etc and copy the appropriate file from: fstab.rm fstab.rmos fstab.rmso fstab.rago fstab.raso fstab.ra fstab.rpos fstab.rp fstab.rko? fstab.up m ( mb up drives) fstab.up m ( mb up drives) fstab.hp m ( mb hp drives) fstab.up (other up drives) fstab.hp (other hp drives) to the file /etc/fstab, i.e.: #cd /etc #cp fstab.xxx fstab this will set up the initial information about the usage of disk partitions. listing . instructions for installing fstab. listing . cogito's instructions for installing fstab. with the largest sum, in this case ops . (please note that the ratings, weights, and assigned values are subjective; they were chosen by senn. the other authors would have assigned different values and therefore might have drawn different conclusions.) user satisfaction we found it easier to install a u n i x operat- ing system using cogito t h a n the u n i x ref- erence manuals because with cogito the in- formation presented to the user was tailored to the specific application. for ex- ample, in the disk-definition state, the file '/etc/fstab' must be created. assume that the boot disk is a 'ampex c o m ' connect- ed to ubao at drive . this implies that the disk address is . also assume the user's name is 'pat' and that pat is ready to install 'fstab'. since the instructions given in the manual are generic, the installer must relate the general instructions to the specific appli- cation. listing is a copy of a relevant sec- tion of the manual. listing is a copy of co- gito's instructions. cogito has remembered information the user gave it a half-hour ago: that the user's name is pat and the disk is an ampex # cd /etc * cp fstab.up m junk * vi junk (edit the file. pat. a. add the line vdev/upob::sw::'. b. give the global substitute command ':g/upo/s//up /'. c. save the new contents and quit the editor.) # cat junk » fstab m. cogito has used this information to make its instructions specific. cogito's instructions are personalized, relevant to the user's task, clear and complete. this example illustrates that the manual's instructions are incomplete and rely on pre- vious and implicit knowledge. the instruc- tions are incomplete because the swap parti- tion '/dev/upob', must be added to the file. they rely on previous knowledge because the user must recall that the u n i x stand- alone disk name for a 'ampex m' is 'up'. they rely on implicit knowledge be- cause the user must somehow know that the disk address of the partitions must be changed from ' ' to t. testing testing cogito was difficult. it was impossi- ble to present the system with every possible combination of inputs and evaluate its out- puts. the best test we devised was to let in- tended users try cogito in many hypotheti- cal situations. if the knowledge base was incomplete, then in some situations the ad- vice given to the users should be incorrect. cogito was tested by the computer sys- tems administrator and a professor of the dept. of systems and industrial engineering and the computer systems administrator for the dept. of computer science at the uni- versity. all found cogito's advice to be com- plete and correct. in an effort to bolster our testing, we asked several nonexpert computer users to try the system. their evaluations empha- sized the difficulties they had using cogito and making sense of its queries or output, and the extent to which they were able to fool the system with plausible (but nonsensi- cal) inputs. recently, we gave cogito a real test. be- cause of inadequate glue on the heads of our ra disk, we had to rebuild our oper- ating system from the distribution tapes. cogito helped us. we found a few omissions in cogito's advice, but no mistakes, and completed the task in about hours. it took us three months the first time we built the system, but we were way down on the learning curve then. this experience has reinforced our belief that all expert systems are inadequately test- ed. no quantitative procedures exist for testing expert systems. most tests merely in- volve running a few case studies; they do not exhaust all possibilities. for example, we are confident that cogito works well for small vax systems, but we cannot be sure it will work as well for s or s. we carefully planned our tests and evalua- tions for dcd. because the dcd system was designed for users with little previous expe- rience in data communication, we allowed juniors from our microcomputer class to ai expert • september test and evaluate it. thirty-three student teams (two members per team) participated in testing and evaluation. it is impossible to test all the combina- tions of problems likely to occur in data communication. therefore, only two typical test cases were given. system users were asked to use dcd to: • design a cable that would connect a wyse terminal to a vax computer and to give advice about setting up the wyse ter- minal to make the interconnection work • design a cable that would connect a wyse terminal to the sytek local area net- work and to give advice about setting up the wyse terminal and the sytek port. we specified two major subtasks in the evaluation process. first, we focused on the accuracy of the system. was the advice given by the system sufficient to make the inter- connection work? this assessment was triv- ial since the advice given by the system re- sulted in go/no-go situations. second, the students evaluated the quality of the human- computer interaction. was the system easy to use? these evaluation criteria should be directly proportional to future frequency of use of the system. all the undergraduate students accom- plished the tasks with an average time of minutes, suggesting that the accuracy of the system was %. the students rated the quality of human-computer interaction as "user friendly." as a further test of the dcd system, a selected graduate student tried to connect a wyse terminal to the vax computer and a wyse terminal to the sytek net using only the manufactures' user and reference manuals. he succeeded after . hours—about three times longer than the undergraduate students who established the same connections by consulting the dcd system. after this initial test, the dcd system was used by juniors at the university for five consecutive semesters. this continued us- age proves its usefulness. however, dcd's usefulness is diminishing as our equipment changes, and none of our knowledge engi- neers have experience with ops to update the knowledge base. general-purpose testing the most difficult aspect of expert system design is testing. the traditional testing method is to have a human expert run many sample cases on the expert system. this consumes a lot of the expert's time and does not guarantee finding all mistakes. con- versely, brute-force enumeration of all in- puts is impossible for most systems. there- fore we have developed a general-purpose tool to help debug knowledge bases without the intervention of human experts. we have tie-based technology for nalltalk- o. .ly there's an expert system shell for the smalltalk- ™ m. for those wanting a powerful way to combine rule- technology with true object-oriented programming, ble™ provides the right tools for the task. iumble is an integrated expert system tool that runs tly in the smalltalk- environment. it features both #ard and forward chaining, a modular certainty m, an advanced user interface including a graphical ser, and a complete programmer interface. > bang for the buck! humble costs far less than >arable expert systems. before buying, ask yourself: / expand this tool to fit my particular needs? are the 'es available if i need to change them? will this tool rate with my own programs? can i port the results of my to end user machines easily? at all?" if you answered to any of these questions, then take a good look at ible. iumble is available for the apple macintosh, sun- iun- series, tektronix and series, apollo [ain series, and xerox and series work- ins. [umble runs in any license version of the smalltalk- >tem. knowledge bases can be transferred between any ltalk- version without modification. [humblef • .u* i«xl«r.' for more information, contact the smalltalk- marketing manager xerox special information systems p.o. box pasadena, ca ( ) - circle on reader service card xerox®, smalltalk- , and humble are trademarks of xerox corporation. * acm conference on object-oriented programming systems, languages, and applications, september - , . san diego, california a. terry bahill is a professor of systems engineering at the university of arizona in tucson. he is also a vice president of the ieee systems, man, and cybernetics society and an associate editor of ieee expert. pat harris is a systems engineer. erich senn is head of technical staff with telekuhrs, schweiz, zurich, switzerland. tried to make this tool generic so it can work on any knowledge base, no matter which expert system shell is used. the first component of the tool is a sim- ple spelling checker with a preprocessor that replaces hyphens, underbars, and par- entheses with spaces and removes terms spe- cific to the shell being used, such as legalvals and automaticmenu. the second component of this tool is designed for backward-chain- ing s h e l l s . it collects all terms that appear in the premises (the //parts) of the rules, and flags as potential mistakes those that do not also appear in the conclusions (the then parts) of the rules or in questions or goal statements. it also traces every term appear- ing in a premise to the conclusion of that rule, then to the premise of another rule, and so on u n t i l a goal statement is reached. if it does not reach a goal statement, then a rule is missing. the third component of this tool, which is not yet completed, acts like tieresias in pointing out rules that do not fit the pattern established by the rest of the rules in the knowledge base. for example, when we ran this component on our animal-discrimina- tion knowledge base, it said, "rule is sus- picious, because most rules that mention the animal's fur also ask if it has spots. rule does not." the fourth component comes into play at run time. the firing of each rule is record- ed. rules that never fire and rules that fire for all test cases are probably mistakes and are brought to the attention of the human expert. size cogito contains rules, kb, and took hours to build. dcd contains rules, kb (plus an interface program and a large help file), and took hours to build. to better understand these numbers, we performed statistical analysis of simple expert systems written by students in partial f u l f i l l m e n t of the requirements of a course in expert systems in . (we call these sys- tems simple because they did not use exter- nal functions or special hardware.) the average student spent hours on the project: % of this time was spent learning the subject, % interviewing the expert, % developing and debugging the knowledge base, and % testing the sys- tem. the average knowledge base con- tained rules and required kb. from these systems and from a less formal evalua- t i o n of other student-generated expert systems constructed over a three-year time span, we concluded that the number of rules (or knowledge base entries) is not a good in- dication of the complexity of a system. however, we also concluded that in rou- tine expert systems, novice knowledge engi- neers consume three to four hours per kilo- byte of knowledge base. sophisticated knowledge engineers may expend as much as five to hours per kilobyte, because they know how to collapse similar rules into more general expressions and might be us- ing external functions. appropriateness we used m.i and ops . one is primarily a backward chainer; the other primarily a for- ward chainer. we do not think much differ- ence exists between the two. for our prob- lems, we could have made either type of chaining work. after learning each of these expert system shells, senn remarked, "it's just another language." this remark has another implication about the knowledge-transfer process. ex- pert systems have made one step forward accompanied, unfortunately, by one step backward. while expert systems permit a higher level of abstraction, they also de- mand that the knowledge base remains tightly coupled to the inference engine. even though the knowledge has already been captured in cogito or dcd, convert- ing this knowledge to another inference en- gine would be difficult. a fundamental breakthrough occurred when traditional programming languages became indepen- dent of the processor; a similar break- through must be effected within the expert system realm. how can one identify a task appropriate for a pc-based expert system? first, there must be a human who performs that task better than most other people. second, the task's solution must be explainable by the human expert in words without relying on pictures. third, the problem must be solv- able a -minute or even one-hour tele- phone conversation with the expert. if it would take a human two days to solve the problem, it is far too complicated for an ex- pert system; if the human gives the answer in two seconds, it is too simple. fourth, problems that involve determining one of many possible solutions are ideal candidates for expert systems, such as answering the question, "what disease does the person have?" fifth, the problem must encompass uncertainty or inexactness in the data, or a diagram on paper would be superior. the first four criteria help predict wheth- er the expert system will be successful, whereas the fifth helps decide whether a complex expert system shell is needed or if some simpler tool would suffice. the animal classification system (figure ) is an example of a problem domain that has no uncertain- ty or inexactness. this task could be per- formed better using a diagram such as fig- ure than an expert system on a pc. given these criteria, giving advice for ai expert • september bringing up u n i x on a vax computer was an inappropriate task for a pc-based expert system. it cannot be done in a one-hour conversation with an expert. we estimate that it would take an expert one or two days to do this task (it took us three months to do it the first time!). cogito's rules filled two floppy disks. we succeeded in making an expert system that worked, but it was hard work. a more powerful tool (such as kee [intellicorp], s.i [teknowlege], art [inference corp.], or knowledge craft [carnegie group inc.]) would have been more appropriate. however, helping a person connect a ter- minal to a computer is an appropriate task for an expert system. in the future, we think most computer equipment manufacturers will have expert systems (interactive text- books) to help their sales personnel install their equipment. this w i l l insulate their re- search engineers from mundane questions and build confidence in the sales force, which will then be able to answer difficult questions about their product. the two real-world expert systems dis- cussed here were designed to perform cer- tain tasks, and they did them; so they were successful. however, the more important as- pect of these endeavors was that we learned a great deal about making expert systems. we learned that an expert system shell is not merlin's long-lost magic wand. you can- not do things with expert system shells that you could not do in any other language; but you can develop a prototype a lot faster. we learned that expert systems can be made friendlier than conventional computer pro- grams. but we also learned that testing and validating expert systems is d i f f i c u l t . q] this research was partially supported by grant no. afosr- - from 'the air force office of scientific research. references . i'xix programmer's manual reference guide. berke- ley, c a l i f . : u s e n i x assoc., . . u\ix systems manager's manual. berkeley, calif.: u s e n i x assoc., . . moller, r., a.t. bahill, and l. swisher. "the de- velopment of esi ac, an expert system for i d e n t i f y i n g a u t i s t i c children," in proceedings of the international conference on systems, man, and cybernetics. new york, n.y.: ieee, , pp. - . . forman, e.h., and t.l. saaty. expert choice. mclean, va.: decision support software, . . saaty, t.l., and k.p. kerns. analytic planning: or- ganization of systems. new york, n.y.: pergamon press, . . harmon, p., and d. king. expert systems: artificial intelligence in business. new york, n.y.: wiley, . . edwards, w. "how to use m u l t i a t t r i b u t e u t i l i t y measurement for social decision making." ieee tran- scripts on systems, man, and cybernetics smc- : - , . . buchanan, b.c., and e.h. s h o r t l i f f e . rule-based expert systems. reading, mass.: addison-wesley, . now available in turbo c,- microsoft c,® jpi modula ,* and logitech modula / turbo expert. now it doesn't take a genius to plug into expert systems. or only $ . , you can incorporate the power of a full-fledged expert system into your turbo pascal programs. seamlessly. a f f o r d a b l y . anally. actual expert systems, developed for simple use by any turbo pascal . programmer. take a look at all the features you suddenly have available with this single turbo pascal . unit: the ability to create large expert systems, or even l i n k m u l t i p l e expert systems together. a powerful backward-chaining inference engine. easy flow of both data and program :ontrol between turbo expert and the other parts of your program, to provide expert system analysis of any database, spreadsheet, file >r data structure. the ability to add new rules in the middle of a consultation, so your expert systems can l e a r n — r e a l l y learn—and )ecome even more intelligent. you also have the ability to create large rule bases and still have plenty of room left for your program, thanks to conservative memory ise. you can link multiple rule bases, you'll be compatible with our turbo companion units, and you have available advanced features like late and time arithmetic, confidence factors, windowing, demons, agendas, blackboards, and more. imagine a single "exe" file containing your user interface and data handling, and a f u l l expert system. for a limited time, get a free copy of our turbo snapshot graphics package worth $ . . we'll give one away with every copy of urbo expert sold between now and september . this package will let you capture graphics images from other programs nd use them in any turbo pascal program. you can even convert images from any cga or ega format to any other. * on top of all that, turbo snapshot has routines for graphic gauges and dials as well as mouse support. you'll have all ou need for a sophisticated graphics f r o n t - e n d for your expert systems — f r e e . b.-.^^.i,.b call for more i n f o r m a t i o n or to order, ( ) - . software artistry inc., depauw blvd., suite , indianapolis, vt.vjj v . i n c l u d e $ . for shipping and handling. circle on reader service card mimicry embedding for advanced neural network training of d biomedical micrographs mimicry embedding for advanced neural network training of d biomedical micrographs artur yakimovich a, �, moona huttunen a, jerzy samolej a, barbara clough b, nagisa yoshida b, c, d, serge mostowy c, d, eva frickel b, and jason mercer a, � a mrc-laboratory for molecular cell biology, university college london, gower st, kings cross, london wc e b, united kingdom b host-toxoplasma interaction laboratory, the francis crick institute, midland rd, london nw st, united kingdom c department of infection biology, london school of hygiene & tropical medicine, keppel street, london wc e ht, united kingdom d section of microbiology, mrc centre for molecular bacteriology and infection, imperial college london, london sw az, united kingdom the use of deep neural networks (dnns) for analysis of com- plex biomedical images shows great promise but is hampered by a lack of large verified datasets for rapid network evolu- tion. here we present a novel “mimicry embedding” strategy for rapid application of neural network architecture-based analysis of biomedical imaging datasets. embedding of a novel biolog- ical dataset, such that it mimics a verified dataset, enables ef- ficient deep learning and seamless architecture switching. we apply this strategy across various microbiological phenotypes; from super-resolved viruses to in vivo parasitic infections. we demonstrate that mimicry embedding enables efficient and ac- curate analysis of three-dimensional microscopy datasets. the results suggest that transfer learning from pre-trained network data may be a powerful general strategy for analysis of hetero- geneous biomedical imaging datasets. deep learning | capsule networks | transfer learning | super-resolution mi- croscopy | vaccinia virus | toxoplasma gondii | zebrafish correspondence: jason.mercer@ucl.ac.uk, artur.yakimovich@ucl.ac.uk introduction artificial neural networks (ann) excel at a plethora of pat- tern recognition tasks ranging from natural language pro- cessing ( ) and facial recognition ( ) to self-driving vehicles ( , ). in biology, recent advances in machine learning and deep learning ( – ) are revolutionizing genome sequenc- ing alignment ( ), chemical synthesis ( , ) and biomedi- cal image analysis ( – ). in the field of computer vision, convolutional neural networks (cnns) perform object detec- tion and image classification at a level matching or surpassing human analysts ( ). despite this, cnn-based architectures often poorly recognise unseen or transformed (e.g. rotated) data due to the use of max or average pooling ( ). while pooling allows cnns to generalize heterogenous data, po- sitional information is ignored. this leads to prioritization of smaller image features and results in an inability of the network to “see the big picture”. to circumvent this, dy- namically routed capsule-based architectures have been pro- posed ( , ). these architectures are nested allowing for the retention of image feature positional information, and op- timization of cnn performance on images with a larger field of view. however, these architectures remain data-hungry and of- ten perform poorly on small biomedical datasets of high complexity ( ). one major reason for this is the lack of large, balanced well-verified biological datasets ( ), akin to mnist ( ) and imagenet ( ) that allow for rapid al- gorithm evolution. to circumvent this, ann analysis of biomedical images can be aided through transfer learning ( , ). for this, weights of a network trained on one dataset are transferred onto a fully or partially identical untrained network which is then trained on a biomedical dataset of a similar nature ( ). this approach shortens training time and is generally considered to be more efficient than random weights initialization strategies ( , ). here, we describe a novel data embedding strategy we term, ‘mimicry embedding’ that allows researchers to circumvent the need for verified biomedical databases to perform ann analysis. mimicry embedding involves transforming biomed- ical datasets such that they mimic verified non-biomedical datasets thereby allowing for mimicry weights transfer from the latter. by embedding d, fluorescent image-based vac- cinia virus and toxoplasma gondii host-pathogen interaction datasets to mimic grey-scale handwritten digits, we demon- strate that mimicry weights transfer from mnist ( ) allows one to harness the performance of cutting-edge ann archi- tectures (capsnet) for the analysis of biomedical data. fur- thermore, the high accuracy of the embedded datasets may allow for their use as novel verified biomedical databases. results more often than not host-pathogen biomedical datasets are not large enough for deep learning. however, we rea- soned that advances in high-content fluorescence imaging ( ) which allow for -d, multi-position single-pathogen res- olution can serve to increase the size of datasets for ann analysis ( ). to classify single-pathogen data in d biomed- ical images we developed ‘zedmate’, an imagej-fiji ( ) plugin that uses the laplacian of gaussian spot detection engine of trackmate ( ). we challenged zedmate with multi-channel, d fluorescent images of late timepoint vac- cinia virus (vacv) infected cells (fig. a and fig. ). owing to its large size, well-defined structure and multi- ple layers of resident proteins that distinguish different virus forms, vacv has the features needed for complex fluores- yakimovich et al. | biorχiv | october , | – certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / cence microscopy-based biomedical particle analysis (fig. b). by detecting and linking individual virions within an image across the z-dimension, zedmate transforms a series of d images into a d dataset (fig. b). from the original four-fluorescent channel composite zed- mate generates grayscale images that preserve the intensity distribution across the z-dimension of each detected channel (fig. c, upper). from this, fluorescence intensity matrices of each channel per z-plane are then generated for individ- ual particles (fig. c; lower). using these matrices and ac- counting for the d positional information of the detected particles, zedmate reconstructions can be plotted (fig. d). intensity analysis across all channels allows for binning of virions into three categories consistent with their biological readouts (fig. b). initial reconstructions indicated that zedmate cannot distin- guish between incoming cell-free virions and newly repli- cated cell-associated virions based solely on c and c in- tensities (fig. d and s ). to improve the precision of zedmate-based binning we devised a binary ml/dl strat- egy relying on manual annotation to separate cell-free from cell-associated virions. to maintain the spatial information acquired in zedmate we attempted to train the capsule ann (capsnet) ( ) on this annotated dataset. these initial at- tempts failed likely due to the small size and complexity of the dataset, two things capsnet struggles with ( ). to circumvent these issues we decided to harness the state- of-the-art performance of capsnet on the relatively simple grayscale dataset, mnist ( ). to generate weights match- ing our binary classification problem, the handwritten digits in mnist were separated into two classes: < and ≤ . with no changes to capsnet other than restricting its output to two classes, this network converged with . % accuracy (fig. a). to allow for transfer learning from this network to our biomedical dataset we designed a vector embedding strat- egy we term “mimicry embedding”. for this, the tensors of each virion’s multi-channel, fluorescence z-profiles from zedmate are assembled across the x-axis. this is followed by linear interpolation and padding which serve to centre the virion in a x pixel image such that the resulting data mimics the grayscale mnist dataset (fig. b). with this approach we aimed to preserve the weights of early cap- snet layers by maintaining the binary mnist capsnet ar- chitecture and performing weights transfer. training on our mimicry-embedded real-world dataset achieved . % accu- racy ( . % precision, . % recall) at separating cell-free from cell-associated virions (fig. b and a-d for classifier training). the capsnet generator was used to visualize how the trained ann distinguished between cell-free and cell-associated virions with such accuracy. the reconstructions indicated that cell-free virions were elongated with moderate intensity profiles while cell-associated virions were compact and very bright (fig. c). the reconstructions were in agreement with mimicry embedded virions suggesting that these properties yielded the base for the high classification accuracy (fig. d). fig. . zedmate facilitates detection and classification of vacv particles in infected cells. (see also fig. ). a, merged four channel fluorescent image of a hela cell infected with vacv (see fig. a for channel details). scale bar; µm. b, illustration of laplacian of gaussian (log)-based vacv particle detection in d. dumbbell shape (red) represents a particle sliced in optical z-sections (semi- transparent grey) providing point signal for log detection (yellow) and connected in z (not to scale). c, intensity measurement from detected particles represented as a z-profile intensity matrix. d, d plot of detected particles color-coded according to detected channels and virion category (see fig. b for details). quantification of different particle types is inset. n= cells, error bars; + sem. | biorχiv yakimovich et al. | mimicry embedding certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / fig. . mimicry embedding allows for separation of cell-free and cell-associated vacv particles through weights transfer from a binary mnist dataset. a, capsnet architecture for training on the mnist hand-written digits dataset repurposed into a binary classification problem (< or ≤ ) prior to capsnet weights transfer. b, mimicry embedding of vacv z-profiles detected by zedmate. the intensity matrix of fluorescence signal (see fig. ) was embedded to mimic mnist data using linear interpolation and padding scale bar; µm. capsnet architecture - with pre-trained weights from a – for training on mimicry embedded vacv particles. c, reconstructed particle profiles of the virions separated as cell-free and cell-associated by capsnet. d, representative mimicry embedded vacv particles for comparison to c. yakimovich et al. | mimicry embedding biorχiv | certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / to verify this strategy, we performed inference on an un- seen (separate from training and validation sets) experimen- tal dataset. fig. shows the workflow from an input four- channel image ( fig. a and s a,b), to detection and binning of virions ( fig. b), followed by mimicry embedding and capsnet separation of cell-free versus cell-associated virions ( fig. c). the results indicate that our model allows for ac- curate classification of virions into four biologically relevant classes within unseen datasets ( fig. d). we’ve established that mimicry embedding and weights transfer allows us to distinguish between incoming cell-free and newly assembled cell-associated virions at late time- points after infection. next, we asked if this approach could also be used to classify extracellular versus intracellular viri- ons during virus entry, a single-cell assay that often requires specific antibodies or labelling strategies and labour-intensive manual annotation. considering these common limitations, we generated a training dataset that would allow for gener- alization of this approach. early infected cells, virions seen in c , were stained with common fluorescent dna (c ) and actin (c ) dyes. to circumvent hand-labelling of the train- ing data, immunolabelling to distinguish between intra- and extracellular virus (c ) was used as a weak labelling ( ) strategy (fig. a and s a). after zedmate detection and transformation of individual particles, intra- and extracellu- lar virus weak labelling (c ) was removed for mimicry em- bedding. by maintaining our binary mnist capsnet archi- tecture and performing weights transfer, we could achieve % accuracy ( . % precision, . % recall) in differentiat- ing between intra- and extracellular virions in the absence of specific-antibody labelling and manual annotation (fig. b-e for classifier training). to estimate accuracy, inference was performed on an unseen dataset in which intra- and extracellular virions were quanti- fied using c -c (measured)- inclusive of extracellular virion weak labelling – or only c -c (predicted) (fig. b). a % match between measured and predicted quantification of in- tracellular particles was seen (fig. b; inset). this indicates that weak labelling can effectively substitute for manual an- notation of training datasets when classifying intra- and ex- tracellular virion signals. as an additional test of the ann, we generated a dataset skewed for extracellular virions by blocking virus entry with ipa- ( , ) (fig. c). consistent with its performance (fig. b-e), a % match between mea- sured and predicted quantifications of intracellular particles was seen (fig. d). finally, when we visualized the recon- structions of intra- and extra- cellular virion classes, extracel- lular virions appeared brighter and more elongated in the z- direction than intracellular ones (fig. e). this was in agree- ment with their mimicry embedded counterparts (fig. f), explaining the anns ability to accurately predict between and quantify these two virion classes. to assess the general applicability of our mimicry embed- ding approach, we acquired a biomedical imaging dataset of cells infected with an egfp-expressing version of the para- site toxoplasma gondii (tg-egfp). while tg-egfp is readily visualized by conventional mi- fig. . inference demonstrates that mimicry embedding and trained cap- snet allows for efficient classification of vacv particles into four biological classes. a, merged four channel fluorescent image of a hela cell infected with vacv previously unseen by capsnet (see fig. a for channel details). scale bar; µm. b, respective zedmate particle detection and classification by conventional binning of fluorescence intensities. c, respective inference of cell-free and cell- associated particles detected by zedmate, mimicry embedded and predicted by a trained capsnet (see fig. b,c). d, combined zedmate particle detection with mimicry embedded and trained capsnet results in classification of four types of bi- ologically relevant vacv particles. inset contains quantification of the particle types in the respective image. | biorχiv yakimovich et al. | mimicry embedding certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / croscopy, detecting and quantifying intracellular viability at the single parasite level is challenging ( ). to generate a tg viability training dataset, cells infected with tg-egfp (c ) were fixed and stained with fluorescent markers of dna (c ), and host cell ubiquitin (c ) which was used a weak label to annotate the subset of “unviable” parasites ( , ) (fig. a). individual particle detection and transformation in zed- mate was followed by mimicry embedding in the absence of c weak labelling. after weights transfer from the binary mnist capsnet architecture (illustrated in fig. b), and fine tuning on the tg-egfp we achieved % accuracy (precision . %, recall . %) in the absence of specific viability la- belling (fig. a-d for classifier training). to assure the ann could accurately distinguish between vi- able and unviable parasites we generated a data set of cells infected with tg-egfp using a specific viability label (c ) as ground truth (fig. a). to further assess viability, experi- ments were performed in the absence or presence of infg, which drives parasite killing. upon model training and val- idation, test inference on this dataset using c -c resulted in a % and % match between measured (c ) and predicted (c -c ) viability in the absence or presence of infg, respec- tively (fig. b). capsnet generator reconstructions showed that “viable” tg-egfp appear larger and brighter than “un- viable” parasites in both c and c (fig. c). this likely ex- plains the ability of the model to accurately predict tg-egfp viability in the absence of specific c viability labelling. in an attempt to train a general model for in vivo parasite vi- ability assessment using our in vitro dataset, we performed mimicry embedding on tg-egfp (fig. a; c ). this resulted in a > % drop in prediction accuracy when training on cap- snet or using autokeras ( )), a neural architecture search (data not shown). this suggested that single channel mimicry embedding does not provide enough context for training of complex algorithms. however, we reasoned as our mimicry embedding is based on mnist, we could use any algorithm that performs well on this dataset. by switching to drop- connect ( ) architecture, which performs among the best on mnist, our classifier achieved % accuracy (precision . %, recall . %) in differentiating between viable and unviable parasites using a single channel (fig. e-h for clas- sifier training). to test this classifier on an in vivo dataset we infected zebrafish (danio rerio) larvae with tg-egfp and imaged them at , and h after infection by fluorescent d- stereomicroscopy (fig. d). zedmate was used to detect and quantify tg-egfp numbers over time (fig. e). a dra- matic drop-off in parasite count was seen between infection at h and h, followed by increased numbers of tg-egfp by h. next the tg-egfp z-profiles were mimicry embed- ded, normalized and their viability inferred using the in vitro infected cell model previously trained on dropconnect. at high pathogen load ( h) % of tg-egfp were scored as viable (fig. f). by h this increased to % without any significant change within h. these results are consistent with an initial clearing of unviable parasites, and replication of the remaining viable ones ( ). fig. . mimicry embedding can be used for weak-labelling particle classifica- tion. a, merged four channel fluorescent image of a hela cell infected with vacv previously unseen by capsnet (see fig. a for channel details). b, zedmate detec- tion and trained capsnet predicted extracellular and intracellular particles. quan- tification of intracellular particles is inset. c, merged four channel image of hela cell infected with vacv and treated with the entry inhibitor, ipa , previously un- seen by capsnet. d, zedmate detection and trained capsnet of intracellular and extracellular particles. quantification of intracellular particles is inset. e, represen- tative reconstruction profiles of extra- and intra- cellular virions. f, representative mimicry embedded extra- and intra- cellular vacv particles for comparison to e. n= untreated and treated cells each. scale bars a-d; µm. yakimovich et al. | mimicry embedding biorχiv | certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / fig. . mimicry embedding and weight transfer employed for toxoplasma gondii (tg) viability detection in cell culture and in vivo.. a, merged three channel fluorescent image of a huvec cells infected tg egfp. individual channels represent dna stain (c ), tg egfp (c ), ubiquitin (c ). scale bar; µm. b, quantification of weak-labelled (measured) and capsnet inferred (predicted) viable and unviable parasites. c, representative reconstructions of the trained capsnet network for viable and non-viable classes of tg egfp z-profiles. d, representative images (maximum intensity projections) of zebrafish (d. rerio) larvae infected with tg egfp at , and h pi. scale bar; µm. e, zedmate detected tg counts at , and h pi f, in vivo inference of tg egfp viability over time using dropconnect viability model trained on in vitro tg data. n= images, error bars + sem. discussion ann analysis of biomedical datasets has trailed behind the unprecedented advancement of ai analysis seen in other fields. this is largely due to the lack of open source, ver- ified biomedical datasets comparable to mnist and ima- genet ( , ). here we present zedmate and mimicry em- bedding as a strategy to harness the power of datasets like mnist and transfer learning to train highly accurate mod- els for analysis of -d biomedical data. zedmate, an open source (imagej/fiji) plugin designed for rapid detection and batch-quantification of d images at the single spot level made mimicry embedding possible. when used together with capsnet ( ) mimicry embedding proved to be a promising method for detection of complex biomedical phenotypes in vitro. we show that transforming real-world images such that they resemble landmark datasets assures compatibility with, and seamless switching between, cutting-edge architectures. embedding data in such a way allows one to maintain full compatibility with weights of the first layers thereby improving transfer. using in vivo biomedical data, we further demonstrate that mimicry em- bedding can yield a model with higher accuracy than one ob- tained through cutting-edge neural architecture search. thus, mimicry embedding can serve as a common denominator for assessing performance between architectures. collectively, our results suggest that zedmate and mimicry embedding, although employed here for the analysis of host-pathogen in- teraction, can be used for ai analysis of any biomedical -d dataset. | biorχiv yakimovich et al. | mimicry embedding certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / materials and methods cell culture, antibodies and reagents. hela cells (atcc) were maintained in in dulbecco’s modified eagle’s medium (dmem, gibco, life technologies, switzerland) with the addition of % fetal bovine serum (fbs, sigma), and % penicillin-streptomycin (pen-strep, sigma), mm gluta- max (life technologies). human umbilical vein endothe- lial cells, huvecs, (c , promocell), were maintained in m medium (gibco) supplemented with mg/ml en- dothelial cell growth supplement (ecgs, – , upstate), units/ml heparin (h- , sigma) and % fbs (sigma). cells were cultivated on plates, pre-coated with % (w/v) porcine gelatin (g , sigma). both huvecs and hela were grown as monolayers at . ◦c and . % co . hu- vec were not used beyond passage . hoechst (sigma) was used post fixation at : , dilution throughout. cell culture grade dimethyl sulfoxide (dmso), used to dissolve control experimental compounds was obtained from sigma. vacv and parapoxvirus strains and virus purification. vac- cinia virus strain western reserve expressing a mcherry protein (vacv wr) was used throughout ( , , ). vacv mature virions (mvs) were purified from cytoplas- mic lysates by being pelleted through a % sucrose cushion for min at , × g. the virus pellet was resuspended in mm tris (ph . ) and subsequently banded on a to % sucrose gradient at , × g for min. following centrifugation, the viral band was collected by aspiration and concentrated by pelleting at , × g for min. mvs were resuspended in mm tris (ph . ), and the titter was deter- mined for pfu per millilitre as previously described ( ). early vacv infection and extracellular virions stain- ing. hela cells were seeded onto cellview slide (greiner bio-one) at , cells per well h before the experiment. vacv a -mcherry f -egfp was added at moi , to in- crease the chances of synchronous infection. cells were fixed with % em- grade pfa hours post infection (hpi) for min followed by a pbs wash. staining and labelling was preceded by blocking (without permeabilization) in block- ing buffer ( % bsa, % fbs, in pbs) for min at room temperature (rt). next, l mouse ( d ) antibody ( ) ( : ) in blocking buffer was added for min at rt, fol- lowed by a pbs wash. anti-mouse antibody (alexa , invitrogen. : ), phalloidin (sigma, : ) and hoechst in blocking buffer were added for min at rt, fol- lowed by a pbs wash. , ’-disulfanediyldinaphthalen- -ol vacv entry inhibitor (ipa- ) was obtained and used as de- scribed ( ). dmso concentration was equal to or below %. late vacv infection and staining. hela cells we cultured on the coverslips and infected with vacv wr expressing a mcherry protein. at hpi cells were fixed with % v/v fa. next, vacv b protein antibody (mouse, : ) in block- ing buffer was added for min at rt, followed by a pbs wash. anti-mouse antibody (alexa ), hoechst in block- ing buffer were added for min at rt, followed by a pbs wash. toxoplasma gondii (tg) cultivation learning infection phenotypes. toxoplasma (rh type i and prugniaud type ii strains) expressing gfp/luciferase were maintained in vitro by serial passage on human foreskin fibroblasts (hffs) cul- tures (atcc). cultures were grown in dmem high glucose (life technologies) supplemented with % fbs (life tech- nologies) at ◦c in % co . tg cultured cells infection and staining. the day be- fore the infection, type ii parasites were passaged onto new hffs to obtain parasites with a high viability. tg were pre- pared from freshly g syringe-lysed hff cultures. para- sites were subsequently x g syringe lysed and excess hff cell debris removed by centrifugation. then, the par- asites were added to the experimental cells at an moi= . the cell cultures with added tg were then centrifuged at x g for min to synchronize the infection and the cultures incubated at ◦c in % co for h. samples treated with interferon gamma (ifnγ) were subjected to iu/ml human ifnγ ( -if, rd systems) for h prior to infection. upon fixation cells were stained with hoechst and mouse mab anti-ubiquitin fk (pw , enzo lifesciences; rrid: ab ) and alexa fluor - conjugated secondary goat anti-mouse (a- , invitrogen; rrid:ab ). tg infection in vivo. tg egfp parasites (type ) were pre- pared from freshly g syringe-lysed hff cultures in % fbs. parasites were subsequently g syringe-lysed and ex- cess hff material removed by centrifugation. after wash- ing with pbs, toxoplasma tachyzoites were resuspended at . x tachyzoites/µl in pbs. larvae were anesthetized with µg/ml tricaine (sigma- aldrich) during the injection procedures and for all live in vivo imaging. all experiments were carried out on tranac background larvae to minimize obstruction of fluorescence signal by pigmentation. dpf larvae were anesthetized and injected with . nl of parasite suspension into the hindbrain ventricle (hbv) as previously described ( ). infected larvae were transferred into individual wells containing . x e me- dia supplemented with methylene blue pre-warmed to ◦c. zebrafish husbandry and maintenance. fish were main- tained at . ◦c on a hr light, hr dark cycle. embryos obtained by natural spawning were maintained in x e me- dia supplemented with . µg/ml methylene blue. ethics statement. animal experiments were performed ac- cording to the animals (scientific procedures) act and approved by the home office (project licenses: ppl p a and p e e c). all experiments were con- ducted up to days post fertilisation. yakimovich et al. | mimicry embedding biorχiv | certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / super-resolution imaging of vacv intracellular viri- ons. supper-resolution microscopy was performed using a x oil immersion objective (na . ) on a vt-isim mi- croscope (visitech; nikon eclipse ti), using nm, nm, nm, nm laser frequencies for excitation. high-content tg egfp imaging in cells. black plastic flat-bottom -well plates (falcon ) were imaged on an opera phenix high content imaging platform using x magnification, z-slices ( . µm/slice) and multiple fields of view per well. images were as single channel -bit tiff files and further processed for zedmate analysis. d tg egfp imaging in vivo. progress of the in vivo infec- tion was monitored by fluorescent stereomicroscopy (leica m fa, leica microsystems, nussloch gmbh, nussloch, germany) at regular time points. all images were obtained with a x objective, at x magnification ( . µm/px) z planes were captured covering a total distance of µm ( . µm intervals). data processing and deep neural network training. our training hardware was based on a single nvidia ti gpu set up in intel core i k sys- tem equipped with gb of ram and an ssd. in- stallation consisted of anaconda python, keras-gpu . , tensorflow-gpu . and knime . . . some models were trained on macbook pro equipped with intel core i cpu using keras . cpu. source code is avail- able under https://github.com/ayakimovich/ zedmate, example dataset under https://github. com/ayakimovich/virus-mnist. further materials are available upon request. references . ronan collobert and jason weston. a unified architecture for natural language processing: deep neural networks with multitask learning. in proceedings of the th international conference on machine learning, pages – . acm. isbn . . florian schroff, dmitry kalenichenko, and james philbin. facenet: a unified embedding for face recognition and clustering. in proceedings of the ieee conference on computer vision and pattern recognition, pages – . . jinkyu kim and john canny. interpretable learning for self-driving cars by visualizing causal attention. in proceedings of the ieee international conference on computer vision, pages – . . sebastian ramos, stefan gehrig, peter pinggera, uwe franke, and carsten rother. detect- ing unexpected obstacles for self-driving cars: fusing deep learning and geometric mod- eling. in ieee intelligent vehicles symposium (iv), pages – . ieee. isbn . . yann lecun and yoshua bengio. convolutional networks for images, speech, and time series. the handbook of brain theory and neural networks, ( ): , . . yann lecun, léon bottou, yoshua bengio, and patrick haffner. gradient-based learning applied to document recognition. proceedings of the ieee, ( ): – , . issn - . . yann lecun, yoshua bengio, and geoffrey hinton. deep learning. nature, ( ): , . issn - . . christof angermueller, tanel pärnamaa, leopold parts, and oliver stegle. deep learning for computational biology. molecular systems biology, ( ): , . issn - . . jennifer n wei, david duvenaud, and alán aspuru-guzik. neural networks for the prediction of organic chemistry reactions. acs central science, ( ): – , . issn - . . marwin hs segler and mark p waller. neuralâsymbolic machine learning for retrosynthesis and reaction prediction. chemistry–a european journal, ( ): – , . issn - . . juan c caicedo, sam cooper, florian heigwer, scott warchal, peng qiu, csaba molnar, aliaksei s vasilevich, joseph d barry, harmanjit singh bansal, and oren kraus. data- analysis strategies for image-based cell profiling. nature methods, ( ): , . issn - . . martin weigert, uwe schmidt, tobias boothe, andreas müller, alexandr dibrov, akanksha jain, benjamin wilhelm, deborah schmidt, coleman broaddus, and siân culley. content- aware image restoration: pushing the limits of fluorescence microscopy. nature methods, ( ): , . issn - . . daniel fisch, artur yakimovich, barbara clough, joseph wright, monique bunyan, michael howell, jason mercer, and eva frickel. defining host–pathogen interactions employing an artificial intelligence workflow. elife, :e , . issn - x. . alex krizhevsky, ilya sutskever, and geoffrey e hinton. imagenet classification with deep convolutional neural networks. in advances in neural information processing systems, pages – . . sara sabour, nicholas frosst, and geoffrey e hinton. dynamic routing between capsules. in advances in neural information processing systems, pages – . . edgar xi, selina bing, and yang jin. capsule network performance on complex data. arxiv preprint arxiv: . , . . rinat mukhometzianov and juan carrillo. capsnet comparative performance evaluation for image classification. arxiv preprint arxiv: . , . . cristian bartolome yiguang zhang and ashwin ramaswami. deepcell: automating cell nuclei detection with. . li deng. the mnist database of handwritten digit images for machine learning research [best of the web]. ieee signal processing magazine, ( ): – , . issn - . . jia deng, wei dong, richard socher, li-jia li, kai li, and li fei-fei. imagenet: a large- scale hierarchical image database. in computer vision and pattern recognition, . cvpr . ieee conference on, pages – . ieee. isbn . . jeremy west, dan ventura, and sean warnick. spring research presentation: a theoret- ical foundation for inductive transfer. brigham young university, college of physical and mathematical sciences, , . . sinno jialin pan and qiang yang. a survey on transfer learning. ieee transactions on knowledge and data engineering, ( ): – , . issn - . . mojca mattiazzi usaj, erin b styles, adrian j verster, helena friesen, charles boone, and brenda j andrews. high-content screening for quantitative cell biology. trends in cell biol- ogy, ( ): – , . issn - . . j. schindelin, i. arganda-carreras, e. frise, v. kaynig, m. longair, t. pietzsch, s. preibisch, c. rueden, s. saalfeld, b. schmid, j. y. tinevez, d. j. white, v. hartenstein, k. eliceiri, p. tomancak, and a. cardona. fiji: an open-source platform for biological-image analysis. nat methods, ( ): – , . issn - . doi: . /nmeth. . . jean-yves tinevez, nick perry, johannes schindelin, genevieve m hoopes, gregory d reynolds, emmanuel laplantine, sebastian y bednarek, spencer l shorte, and kevin w eliceiri. trackmate: an open and extensible platform for single-particle tracking. methods, : – , . issn - . . alexander j ratner, christopher m de sa, sen wu, daniel selsam, and christopher ré. data programming: creating large training sets, quickly. in advances in neural information processing systems, pages – . . sean w deacon, alexander beeser, jami a fukui, ulrike ee rennefahrt, cynthia myers, jonathan chernoff, and jeffrey r peterson. an isoform-selective, small-molecule inhibitor targets the autoregulatory mechanism of p -activated kinase. chemistry biology, ( ): – , . issn - . . j. mercer and a. helenius. vaccinia virus uses macropinocytosis and apoptotic mimicry to enter host cells. science, ( ): – , . issn - (electronic) - (linking). . barbara clough, joseph d wright, pedro m pereira, elizabeth m hirst, ashleigh c johnston, ricardo henriques, and eva-maria frickel. k -linked ubiquitination targets toxoplasma gondii for endo-lysosomal destruction in ifnγ-stimulated human cells. plos pathogens, ( ):e , . issn - . . haifeng jin, qingquan song, and xia hu. auto-keras: efficient neural architecture search with network morphism. arxiv preprint arxiv: . , . . li wan, matthew zeiler, sixin zhang, yann lecun, and rob fergus. regularization of neural networks using dropconnect. icml’ , pp. technical report, iii– –iii– . jmlr. org, . . nagisa yoshida, marie-charlotte domart, artur yakimovich, maria j. mazon-moya, lucy collinson, jason mercer, eva-maria frickel, and serge mostowy. in vivo control of toxo- plasma gondii by zebrafish macrophages. biorxiv, . . j. mercer, s. knebel, f. i. schmidt, j. crouse, c. burkard, and a. helenius. vaccinia virus strains use distinct forms of macropinocytosis for host-cell entry. proc natl acad sci u s a, ( ): – , . issn - (electronic) - (linking). . samuel kilcher and jason mercer. dna virus uncoating. virology, : – , . issn - . . jason mercer and paula traktman. investigation of structural and functional motifs within the vaccinia virus a phosphoprotein, an essential component of the virion membrane. journal of virology, ( ): – , . issn - x. . elizabeth j wolffe, s vijaya, and bernard moss. a myristylated membrane protein encoded by the vaccinia virus l r open reading frame is the target of potent neutralizing monoclonal antibodies. virology, ( ): – , . issn - . . maria j mazon-moya, alexandra r willis, vincenzo torraca, laurent boucontet, avinash r shenoy, emma colucci-guyon, and serge mostowy. septins restrict inflammation and pro- tect zebrafish larvae from shigella infection. plos pathogens, ( ):e , . issn - . | biorχiv yakimovich et al. | mimicry embedding certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://github.com/ayakimovich/zedmate https://github.com/ayakimovich/zedmate https://github.com/ayakimovich/virus-mnist https://github.com/ayakimovich/virus-mnist https://doi.org/ . / appendix . supplementary information this is a supplementary section to the preprint manuscript by yakimovich et al. source code is available under https: //github.com/ayakimovich/zedmate. all further information is available upon request. supplementary figures. figure s . individual channels used in late vacv infected cells and their biological relevance). a, maximum intensity projections of individual channel images of hela cells infected with vacv at hours post infection. here, dna stain (c ), vacv core a -mcherry (c ), vacv outer envelope protein f (c ) and vacv outer envelope protein b (c ). scale bar; µm. b, illustration of the position of markers in virions [mv (mature virions), iev (intracellular enveloped virions), cev (cell-associated extracellular virions)], and these virions - with the corresponding markers - in infected cells. here c marks cellular dna and cytoplasmic vacv replication sites, c marks all virions (mvs, ievs and cevs), c marks a subset of virions (ievs and cevs) and c marks only cevs. yakimovich et al. | mimicry embedding biorχiv | certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://github.com/ayakimovich/zedmate https://github.com/ayakimovich/zedmate https://doi.org/ . / figure s . capsnet training and validation of cell-free vs. cell-associated virus model. a, late model loss function change upon training iterations (epochs). b, late model training and validation (unseen data) accuracy change upon training iterations (epochs). c, late model receiver operational characteristics (roc) curve of the trained model obtained using unseen data (validation). here area under the curve (auc) was . . d, late model confusion matrix of the trained model obtained using unseen data (validation). late model precision was . %, recall was . %, f -score was . %. | biorχiv yakimovich et al. | mimicry embedding certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / figure s . capsnet training and validation of extracellular vs. intracellular virus model. a, maximum intensity pro- jections of individual channels from hela cells infected with vacv. here, dna stain (c ), vacv core a -egfp (c ), actin stained with phalloidin (c ) and vacv membrane protein l as an extracellular virion label (c ) b, model loss function change upon training iterations (epochs). c, model training and validation (unseen data) accuracy change upon training iterations (epochs). d, model receiver operational characteristics (roc) curve of the trained model obtained using unseen data (valida- tion). here area under the curve (auc) was . . e, model confusion matrix of the trained model obtained using unseen data (validation). model precision was . %, recall was . %, f -score was . %. yakimovich et al. | mimicry embedding biorχiv | certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / figure s . capsnet and dropconnect training and validation of in vitro and in vivo tg egfp viability model. a, -channel capsnet model loss function change upon training iterations (epochs). b, -channel capsnet model training and validation (unseen data) accuracy change upon training iterations (epochs). c, -channel capsnet model receiver operational characteristics (roc) curve of the trained model obtained using unseen data (validation). here area under the curve (auc) was . . d, -channel capsnet model confusion matrix of the trained model obtained using unseen data (validation). e, - channel dropconnect model loss function change upon training iterations (epochs). f, -channel dropconnect model training and validation (unseen data) accuracy change upon training iterations (epochs). g, -channel dropconnect model receiver operational characteristics (roc) curve of the trained model obtained using unseen data (validation). here area under the curve (auc) was . . g, -channel dropconnect model confusion matrix of the trained model obtained using unseen data (validation). the -channel capsnet model precision was . %, recall was . %, f -score was . %. the -channel dropconnect model precision was . %, recall was . %, f -score was . %. | biorχiv yakimovich et al. | mimicry embedding certified by peer review) is the author/funder. all rights reserved. no reuse allowed without permission. the copyright holder for this preprint (which was notthis version posted october , . ; https://doi.org/ . / doi: biorxiv preprint https://doi.org/ . / supplementary information nsf workshop, washington, dc may - , - about about agenda workshop organizers keynote speakers recommended reading about about welcome to the webpage of the nsf operations engineering (oe) program workshop on decision analytics for dynamic policing. this workshop is funded by nsf (nsf cmmi- ) to explore research directions in decision analytics and operations research (da/or) for enabling improved dynamic policing strategies from a variety of aspects. nsf program directors will be welcome to attend at any time during the workshop. in these pages, you will find general information about the workshop (below) and information on workshop agenda, workshop organizers, keynote speakers, and recommended reading. all workshop attendees should review the university of texas at arlington’s code-of-conduct policy. the workshop code of conduct with follow this policy. all workshop attendees should review the university of texas at arlington’s code-of-conduct policy. the workshop code of conduct with follow this policy. https://www.uta.edu/policy/hop/ - date and location date and location date: may - , location: renaissance arlington capital view hotel, arlington, va project summary and report nsf workshop report (may , ) the traditional approach in law enforcement studies utilizes randomized experimental design. however, this approach does not take advantage of today’s data collection capabilities. within the last decade, the concept of predictive policing was proposed by a rand report, funded by the national institute of justice (nij), to categorize and encourage academic research developing machine learning and statistical modeling tools to accurately predict crime. unfortunately, existing predictive policing methods are flawed in four critical aspects: ( ) they omit the human perspective, including human behavior, political biases, and community perception.  ( ) they assume that the data include the necessary information for both prediction and action. ( ) they assume that historical data on the policing system will be representative for the future. ( ) they assume unlimited police resources. a key element in the rand report is the concept of prediction-led policing. for this concept, it is recognized that prediction alone is inadequate without connection to a policing strategy that identifies appropriate actions. this proposed workshop on decision analytics for dynamic policing organized local teams of police departments (pds) and academics with expertise in da/or and criminal justice (cj) to build a cross-disciplinary understanding of policing challenges that may be addressable by da/or. discussions took place within and across these populations, so as to examine all three perspectives. da/or as a discipline is not well known within the law enforcement context. towards filling this gap, the workshop identified and recommended new directions for research in da/or to develop methods for policing that are cost-effective, consider social influences, identify the appropriate data, address the dynamic nature of policing, and utilize existing knowledge from pd/cj expertise. overall, the potential impact of da/or research in law enforcement is extremely high, but progress will require care, collaboration, persistence, and an open mind. objectives the objectives of the proposed workshop are: build an understanding of the dynamics of policing systems. build an understanding of the controllable decisions and uncertain elements in dynamic policing systems, for example, intervention teams, criminal responses to police actions, community outreach, interactions between pds and disciplines related to policing (such as forensic science divisions), etc. create processes that facilitate research partnerships between da/or, cj, and pds to encourage the development of da/or methods for dynamic policing systems. identify critical research themes in dynamic policing that may be studied using da/or. identify methodological advancements needed in da/or to address the critical research themes in dynamic policing. program nsf workshop on decision analysis for dynamic policing program (pdf) department of industrial, manufacturing, and systems engineering box | west first street | woolf hall | arlington, tx © the university of texas at arlington | college of engineering -se -zhao_tomej send orders for reprints to reprints@benthamscience.ae the open mechanical engineering journal, , , - - x/ bentham open open access modularized cutting tool selection expert system chen wang , wu zhao,* , ling chen , kai zhang and xin guo school of manufacturing science and engineering, sichuan university, cheng du . china division of production and materials engineering, lund university, lund, , sweden abstract: this paper is to present a rule-based cutting tool selecting expert system which has knowledge modules and rule bases. besides, according to different process targets, the selection progress will apply corresponding constraints and rule modules. the logic of tool selection follows a decision-making procedure as an experienced engineers. the strategy of system is to guide the user through several standard steps: information input; feature recognition; selection of machining method; selection of tool material and type; calculation of process parameter and solving cutting problem. this system also has a modularized structure which allows adding new functions and new modules to expand knowledge base and data base. modules involves in this system are composed of the user interface, knowledge acquisition facility, explanation facility, the knowledge base module, the inference engine and the database module. keywords: tool selection, expert system, modularization, process optimization. . introduction in mechanical manufacture industry, process planning involves scheduling resources, such as machine tools, work piece, cutting tools, operation sequence, processing parameters and the choice of auxiliary functions [ ]. during the last years some process planning expert systems have been developed to select tools and cutting parameters for specific operations. in early stage, process engineers selected adequate cutting parameters with their experience. obviously, this kind of parameters were selected intuitively considering safety factor. in order to computerize and achieve more accurate cutting process solutions, automation techniques were applied. then, engineers tried to apply intelligent techniques to optimize of the cutting parameters. those intelligent techniques includes several optimization methods, such as: genetic algorithms [ ] (cus & milfelner ), neural networks [ ] (abellan, romeros, siller, estrud, & vila, ) and bio-inspired techniques [ ] (zain, a. m., haron, h., & sharif, s. ). however, these methods provide intrinsic mathematical nonlinear function, hidden in a "black box" giving solutions, on the other hand, it decreases the transparency of process. in the field of manufacturing process planning, expert system involves two steps. first of all, knowledge base system extracts knowledge or experience and then applies fuzzy logic to deduce solutions. as commonly discussed in scientific literature, an online expert system which selecting process information applies fuzzy logic as reasoning mechanism to acquire the knowledge of mechanical engineers [ ] (wong and hamouda ). on the other hand, in order to calculate the cutting parameters and manufacturing route in cutting process, a new computer aided process planning is developed [ ] (vidal, a., alberti, m., ciurana, j., & casadesus, m. ). moreover, in order to acquire optimal process parameters subject to specific precision and surface quality, tool life expectancy and production time sunder different processing conditions, some theoretical and experimental work has been carried out by organizations in several countries [ - ] (chien&chou, ; vivanco, luis, costa, & ortiz and references therein). these expert systems mainly dealt with the geometrical matters and selection of cutting tools [ ]. most of them concentrated on specific components and realize part function. some expert systems took the component material and geometric features into consideration, however, ignored the selecting process of the cutting tool type and material. the main aim of this research is to help designers and manufacturing planners to select optimal cutting tools and optimum cutting conditions to reduce cost and time, as well as to solve some technical problems in cutting process with the aid of innovation strategy. the proposed modularized muti-objective cutting process expert system implements the following features, such as component feature recognition; selecting tool material; selecting tool type and optimum cutting conditions and innovative design. components of cpes expert system use a large number specific domain knowledge of experts to solve problem [ ]. and essentially evolves two kind of terms. one type is constructed with specific kinds of programming languages and tools such as rule-based systems, frame-based systems, and programming languages such as prolog and lisp [ ]. another type is more appropriate for expert system, because expert system can reasons problem as human expert [ ]. this paper tries to propose a classical optimization method integrating innovation strategy, and the design of expert system is based modularized cutting tool selection expert system the open mechanical engineering journal, , volume on modularize method. this system also applied theoretical models and semantic method. the proposed expert system involves user interface, inference engine, explanation facility, knowledge acquisition facility, knowledgebase, and working memory and other executable programs. the structure of the proposed expert system is illustrated as fig. ( ). . . inference engine the inference engine which manipulates the stored knowledge to reach solutions is one of the basic components of expert system. the inference engine could search knowledge base, and provide answers to user by applying rules to the solution of a particular problem [ ]. the scanning process of knowledge will continue until these antecedents match the assertions in database, then user will get the deduced results. besides, the interpreter provides explanations and can be updated by adding a new separate knowledge base module. . . modularized knowledge base the knowledge of cutting process expert system mainly comes from two sources: experience of domain experts and technical documents [ ]. owning an expandable structure is important for the expansibility of knowledge base, just like domain experts learns new knowledge and applies this knowledge in the future. an independent and modularized knowledge base is proposed, this makes it possible to update the knowledge base by simply input data and editing the knowledge information. the modularized knowledgebase mainly involves factual statements, frames, objects, cases, tables, if-then rules and equations [ ]. knowledge representation technology and rules are the key components of knowledge base. relevant information about component characteristics and cutting tools in machining process are included in the hierarchy shown as fig. ( ). elements called ‘objects’ were presented as frames in hierarchy model. the framework consists of a set of slot which contains a description for the attribute of object. characteristics of the components, cutting tools and machining process are also expressed as objects which can be classes. the objects relation of the expert system is composed together in a hierarchy building. system identify the characteristics of objects and then apply the appropriate rules to perform the task. rules in each module will deduce a serious of corresponding conclusions according to correlated conditions. this procedure reflects the logic of knowledge base. what we want is a knowledge base with fine structure rather than a huge plant knowledge base which includes all the rules. the relevant knowledge in expert system is represented in the terms of the formal logic. the knowledge base contains a sequence of rules in a modular approach. thus, expert system can be used in modularized way to adapting to different conditions. the proposed knowledge base here includes rules which can be divided into several modules. in each module, rules will be applied according to the proposed approach. these rules can be split into the modules shown as table . . . database database systems mainly involves two main aspects: data and mathematical model [ ]. the cutting tool data about fig. ( ). structure of the proposed system. the open mechanical engineering journal, , volume wang et al. cutting tool materials, machine tools and work piece materials in catalogues are mainly obtained from the several cutting tool manufactures, large factories and handbooks. the recommended data are put into forms in the database. mathematical model can be used to calculate the cutting parameters based on the database. the database of system consists of several independent groups of data, shown as table , involving materials, cutting tools, tool materials, cutting parameters, processing method and equations, respectively. . system logic the cutting process expert system run under an interactive operation environment. users can describe the characteristic of component with the information integrated into the database. in the process of the operating the software, expert system will guide user through these main steps, including: inputting information, feature recognition, selecting machining process; selecting tool material; selecting tool type; optimizing cutting conditions and solving problems. the scenario of cutting process expert system is shown as fig. ( ). . . process information input and feature recognition the detail information about work piece is the foundation of reasoning process. under the current technical level, characters identified by engineers have more efficient and fig. ( ). hierarchy tree of system. table . knowledge modules. no. module function i stability robustness and allowable gives robustness to the system and presents the rough allowable input cutting parameters. ii control of cutting forces keeping the forces of the system under prescribed upper limit in spite of variations in system parameters. iii drives constraints covers the spindle and feed drive motors constraints. iv initial cutting parameter selection gives initial computational input parameters subjected potential initial spindle speed candidates. v cost function definition and rules to inference with it a novel multi-objective cost function to evaluate the performance of the system. vi automatic feedback and monitoring automatic feedback to the system. vii expert human machine interface monitor key parameters to better inter-actuation with expert engineers and operators. types d rills twist drill flat drill deep hole drill reaming drill center drill hollow drill m ill cutters face milling cutter vertical milling side milling cutter angle cutter saw blade milling t cutter e lectrodes graphite electrodes t urning tools boring threading ext. machining param eters cutting speed tool life metal removalrate m achining type milling drilling turning edm working  piece m etals alloy steel aluminum plastics ... d im ensions three high geometrical feature precision roughness ... m aterial properties hardness toughness machinability thermal properties ... status feature holes pockets threading surface and side machining slots tool  material u ltrahard pcbn cbn pcd c eram ics al o mix si n -sialon c em ented carbide wc-c coated carbides h ss m type t type coated c erm ets coated uncoated knowledge  base modularized cutting tool selection expert system the open mechanical engineering journal, , volume accurate compare to characters identified by computer programs. the identification of piece involves feature specification and dimensions, which need to be entered to system by user. this system provide standardized typical characteristics for user to select. in this stage, system also offer d model with typical geometric features for referring. for example, turning surfaces has some of the following characteristics such as face, diameter, arc, recess, grooving and taper, etc. and these characteristics are also be divided as two groups according to the importance, shown as fig. ( ). . . selecting machining process the above information about work piece enables expert system to select the appropriate cutting process and cutting material. the selection machining method need to consider the these factors such as the type, size and precision, surface finishing, dimensional tolerance of the feature. in this stage, system will select the processing methods including fig. ( ). feature recognition. primary features -cylinder -face -taper -concave -convex -chamfer -radius secondary features -nicked radius -groove -recess -thread -axial recess fig. ( ). the scenario of cutting process expert system. the open mechanical engineering journal, , volume wang et al. traditional processing methods such as drilling, turning, milling, etc. and special working technology such as laser, edm, etc. the relationship between the aspect ratio and the machining type is shown as fig. ( ). taking the typical surface finishing process for example, once the feature of curved surface is identified and the parameter is defined, then the system will make judgment. fig. ( ). corresponding processing method to features. . . selecting the cutting tool the tool selection is the progress of information determination about tool holder, insert, cutting conditions, and coolant. cutting tools are mainly composed of two parts: tool holder and indexable insert. the selecting sequence is first to select tool holder and then suiting insert. system offer two approaches for selecting cutting tools, manual searching based on the database and automatic selection by system. manual searching which has more reliability was performed by the user owing abundant of tacit knowledge. the required information for tool selection includes: tool holder (type, clamping system, hand of cut, point angle, size, etc.), insert (size, shape, nose radius, grade, etc.), processing conditions, and coolant. what’s more, system will recommend reference case and relevant parameter to user, according to machinability data, feature. the parameters of cutting tools are included as iso code in table contain the information about the parameters, carbide grades and functions. selecting tool are carried out through a series of if-then structures to choose the suitable tool holder and insert from the tool library. the material of cutting tools relies on the material of component to be machined. so it is needed to select the material of component firstly, then cpes will deduce the corresponding tool material. there are three way for use to select component material in this system. one option is that user selects material from the databases stored in expert system. another way is to browse and select the material from the cambridge engineering selector (ces) linked to expert system. the last option allows users input the material by themselves depending on own experiment and knowledge. besides, this system provides a serious of rules for selecting suitable tool material corresponding to the selected material of component. . . calculating process parameters the system offers extra function to calculate process parameters such as machining cost and time. the data put in by user can be processed through these equations and rules to calculate parameters. mathematical stored as rules in knowledge modules are shown in table . taking the module v for example, this module is designed to calculate cost and includes serious of equations and rules. inference engine can use this cost function to calculate the parameters. for example, the formula ( ) can be applied cylindricalpocketsslots feature types feature specification surfaces emd laser turning drilling milling ... rules parameter   fig. ( ). selection of tool type. modularized cutting tool selection expert system the open mechanical engineering journal, , volume to calculate the tool cost of milling, which contains other subfunction, such as tol: life of the tool; mrr: material remove rate; surf: surface finish; ros: robustness of the system. they are defined as following: j(tol,mrr,tes,r,ci(i= ,... )) = c ⋅nf ⋅tol + c ⋅nf ⋅mrr + c ⋅ nf surf + c ⋅nf ⋅ros ( ) where ci is weighting factor, nf represents normalisation factor. tol = ktol ⋅v −α ⋅αdc −α ⋅ ft −α ( ) where ktol is model constant, v is cutting speed (m/min), αi is model parameter, adc represents axial depth of cut (mm)and ft represents feed per tooth (mm/tooth). mrr = adc ⋅r dc⋅fc ( ) where rdc represents radial depth of cut (mm) and fc represents feed velocity (mm/s). surf = ksurf ⋅v β ⋅ fc β ⋅αdc β ( ) where ksurf is a model constant and, bi represents surface roughness. ros = min sqrt αdc,ros( ) + ss,ros( ) }{( ) ( ) where ss, ros is the length of the spindle speed. nfi = ji − jmax jmax − jmiin ( ) j = ci ⋅nfi ⋅ji j= ∑ ( ) besides, there are some more simple rules to estimate process time and cost. pt = vf rmr ( ) where pt (min) is process time, vf (mm ) is form feature volume, rmr (mm /min) is material removal rate. the equations and rules of estimated cutting time and feature cost are shown as below: set value ((processing: rpmd), . * processing: spd mmmin- /features: di mm); set value ((processing: frdmmin- ); processing: frdmmrev- * processing: rpmd); set value ((processing: ctd min., features: de mm/ machining: frd mmmin- ); set value ((processing: cost m¥), processing: ctm min * ); set value ((processing: cost d¥), processing: ctd min* ); set value ((processing: cost t¥), processing: ctt min* ); if (processing cost: drilling