A Reinforcement Learning-Based Framework for Multi-Criteria Decision-Making
DOI:
https://doi.org/10.59543/ydm6xr12Keywords:
Multi-Criteria decision-Making, Reinforcement Learning, Learning-to-Rank, Preference AggregationAbstract
Multi-Criteria Decision-Making (MCDM) methods have been applied to various ranking problems in finance, engineering, and management; however, many traditional methods rely on data normalization and fixed weights assigned to the criteria. Often, such data normalization has led to distortion of ordinal information and inconsistent results. In this study, we propose an RL-based framework for solving MCDM problems when the decision data are ranked. The ranking task was then reformulated as a learning-to-rank optimization problem, where the preference scores are assigned to alternatives and learned by maximizing agreement with a benchmark ranking. The proposed approach was first illustrated with a simple example and then applied to the real-world evaluation of banks using the CAMELS criteria. The results showed that the RL-based approach achieved greater consistency with benchmark rankings than normalization-based MCDM methods, thereby extending the applicability of rank-preserving approaches.
Downloads
Published
How to Cite
Issue
Section
License
Copyright (c) 2026 Laxminarayan Sahoo, Sumanta Lal Ghosh (Author)

This work is licensed under a Creative Commons Attribution 4.0 International License.
ISRAJ is published Open Access under a Creative Commons CC-BY 4.0 license. Authors retain full copyright, with the first publication right granted to the journal.





