TR2023-146

Preference-Guided Bayesian Optimization for Control Policy Learning: Application to Personalized Plasma Medicine


    •  Shao, K., Romeres, D., Chakrabarty, A., Mesbah, A., "Preference-Guided Bayesian Optimization for Control Policy Learning: Application to Personalized Plasma Medicine", Advances in Neural Information Processing Systems (NeurIPS), December 2023.
      BibTeX TR2023-146 PDF
      • @inproceedings{Shao2023dec,
      • author = {Shao, Ketong and Romeres, Diego and Chakrabarty, Ankush and Mesbah, Ali},
      • title = {Preference-Guided Bayesian Optimization for Control Policy Learning: Application to Personalized Plasma Medicine},
      • booktitle = {Advances in Neural Information Processing Systems (NeurIPS)},
      • year = 2023,
      • month = dec,
      • url = {https://www.merl.com/publications/TR2023-146}
      • }
  • MERL Contacts:
  • Research Areas:

    Control, Machine Learning, Optimization

Abstract:

This paper investigates the adaptation of control policies for personalized dose deliv- ery in plasma medicine using preference-learning based Bayesian optimization. Preference learning empowers users to incorporate their preferences or domain expertise during the exploration of optimal control policies, which often results in fast attainment of personalized treatment outcomes. We establish that, compared to multi-objective Bayesian optimization (BO), preference-guided BO offers statistically faster convergence and computes solutions that better reflect user preferences. Moreover, it enables users to actively provide feedback during the policy search procedure, which helps to focus the search in sub-regions of the search space likely to contain preferred local optima. Our findings highlight the suitability of preference-learning-based BO for adapting control policies in plasma treatments, where both user preferences and swift convergence are of paramount importance.