-1

I have built a custom discrete event simulator with multiple agents and want to optimize the system using RL frameworks that support multi-agent configurations. I will use custom policies. Which framework should I use? I've looked into SB3, CleanRL, Tianshou, SKRL, RLlib, Acme, and MARLlib, and here's what I found:

  1. SB3 and CleanRL don't offer direct support for multi-agent systems.

  2. RLlib is very functional but has a steep learning curve and hard to customize and some say the documentation is poor.

  3. Tianshou seems good, but its community is small.

  4. Acme doesn't use the PyTorch backend, which I prefer.

  5. I haven't delved deeply into SKRL or MARLlib, but they appear promising.

I prioritize ease of use and documentation. What framework do you suggest? And why?

I’d appreciate any helpful starting advice/resource to approach my problem as well.

1 Answers1

1

Disclaimer: I have only experience with the first three frameworks.

Among those three, I would suggest RLlib for the multi-agent situation:

  1. Distributed training based on Ray library
  2. Out-of-box multiagent solution
  3. Great documentation. You may want to check an example how to train multiple agents: https://applied-rl-course.netlify.app/en/module2
penkovsky
  • 264
  • 1
  • 10