That respect these constraints. In order to achieve this: (i) All agents that do not satisfy the constraints are discarded; (ii) for each algorithm, the agent leading to the best performance in average is selected; (iii) we build the list of agents whose performances are not significantly different. This list is obtained by using a paired sampled Z-test with a confidence level of 95 , allowing us to determine when two agents are statistically equivalent (more details in S3 File). The results will help us to identify, for each experiment, the most suitable algorithm(s) depending on the constraints the agents must satisfy. This protocol is an extension of the one presented in [4].4 BBRL libraryBBRL (standing for Benchmaring tools for Bayesian SNDX-275 biological activity Reinforcement Learning) is a C++ opensource library for Bayesian Reinforcement Learning (discrete state/action spaces). This library provides high-level features, while remaining as flexible and documented as possible to address the needs of any researcher of this field. To this end, we developed a complete command-line interface, along with a comprehensive website: https://github.com/mcastron/BBRL. BBRL focuses on the core operations required to apply the comparison benchmark presented in this paper. To do a complete experiment with the BBRL library, follow these five steps: 1. We create a test and a prior distribution. Those distributions are represented by Flat Dirichlet Multinomial distributions (FDM), parameterised by a state space X, an action space U, a vector of GS-9620 custom synthesis parameters , and reward function . For more information about the FDM distributions, check Section 5.2. ./BBRL-DDS –mdp_distrib generation \ –name \ –short_name \ –n_states –n_actions \ –ini_state \ –transition_weights \ <(1)> ???<(nX nU nX)> \ –reward_type “RT_CONSTANT” \ –reward_means \ <(x(1), u(1), x(1))> ???<(x(nX), u(nU), x(nX))> \ –output