next up previous
Next: Problem Assumption 2: How Up: Interpretation of Results and Previous: Interpretation of Results and

Problem Assumption 1: Are General Purpose Planners Biased Toward Particular Problems/Domains?

The set of problems on which a planner was developed can have a strong effect on the performance of the planner. This can be either the effect of unintentional over-specialization or the result of a concerted effort on the part of the developers to optimize their system to solve a specific problem. With one exception, every planner fared better on the tailored subset of problems (training set). Consequently, we must conclude that the choice of a subset of problems may well affect the outcome of any comparison.

A fair planner comparison must account for likely biases in the problem set. Good performance on a certain class of problems does not imply good performance in general. A large performance differential for planners with a targeted problem domain (i.e., do well on their focus problems and poorly on others) may well indicate that the developers have succeeded in optimizing the performance of their planner.

Recommendation 4: Problem sets should be constructed to highlight the designers' expectations about superior performance for their planner, and they should be specific about this selection criteria.
On the other hand, if the goal is to demonstrate across the board performance, then our results at randomly selecting domains suggests that biases can be mitigated.
Recommendation 5: If highlighting performance on ``general'' problems is the goal, then the problem set should be selected randomly from the benchmark domains.

next up previous
Next: Problem Assumption 2: How Up: Interpretation of Results and Previous: Interpretation of Results and
©2002 AI Access Foundation and Morgan Kaufmann Publishers. All rights reserved.