An Approach to the Automatic Comparison of Reference Point-Based Interactive Methods for Multiobjective Optimization

Solving multiobjective optimization problems means finding the best balance among multiple conflicting objectives. This needs preference information from a decision maker who is a domain expert. In interactive methods, the decision maker takes part in an iterative process to learn about the interdep...

Ausführliche Beschreibung

Gespeichert in:
Bibliographische Detailangaben
Veröffentlicht in:IEEE access 2021, Vol.9, p.150037-150048
Hauptverfasser: Podkopaev, Dmitry, Miettinen, Kaisa, Ojalehto, Vesa
Format: Artikel
Sprache:eng
Schlagworte:
Online-Zugang:Volltext
Tags: Tag hinzufügen
Keine Tags, Fügen Sie den ersten Tag hinzu!
Beschreibung
Zusammenfassung:Solving multiobjective optimization problems means finding the best balance among multiple conflicting objectives. This needs preference information from a decision maker who is a domain expert. In interactive methods, the decision maker takes part in an iterative process to learn about the interdependencies and can adjust the preferences. We address the need to compare different interactive multiobjective optimization methods, which is essential when selecting the most suited method for solving a particular problem. We concentrate on a class of interactive methods where a decision maker expresses preference information as reference points, i.e., desirable objective function values. Comparison of interactive methods with human decision makers is not a straightforward process due to cost and reliability issues. The lack of suitable behavioral models hampers creating artificial decision makers for automatic experiments. Few approaches to automating testing have been proposed in the literature; however, none are widely used. As a result, empirical performance studies are scarce for this class of methods despite its popularity among researchers and practitioners. We have developed a new approach to replace a decision maker to automatically compare interactive methods based on reference points or similar preference information. Keeping in mind the lack of suitable human behavioral models, we concentrate on evaluating general performance characteristics. Such an evaluation can partly address the absence of any tests and is appropriate for screening methods before more rigorous testing. We have implemented our approach as a ready-to-use Python module and illustrated it with computational examples.
ISSN:2169-3536
2169-3536
DOI:10.1109/ACCESS.2021.3123432