Autonomous driving cars, human-like NPCs in videogames, nuclear fusion controller, faster matrix multiplication and ChatGPT: Reinforcement Learning (RL) is nowadays ubiquitus. In Orobix we have built a Deep RL Framework that lets anyone learn and play with agents acting in complex environments.
Autonomous driving cars, human-like NPCs in videogames, nuclear fusion controller, faster matrix multiplication, ChatGPT and more: Reinforcement Learning (RL) is nowadays ubiquitus and when combined with sensory inputs coming from different modalities, i.e. language, image, video, sound or even nuclear physics, it is able to obtain super-human results on the task it is asked to resolve. In the last few months we have witnessed a broader recognition of such methods thanks to the huge impact they have made in our day-by-day life: Deepmind controlled a nuclear fusion at EPFL, while OpenAI released ChatGPT, a language model which interacts with the users in a conversational way. Both of these methods was trained with Reinforcement Learning, with the latter asking feedbacks from human.
In Orobix we have built a Deep RL Framework, called sheeprl, that lets anyone learn and play with agents acting in complex environments, focusing in particular on training agents to compete with human players on the MotoGP videogame. The framework has been written completely in Python, leveraging the power of PyTorch and Hydra for the deep learning modules and the configurations respectively.
This talk would mainly cover the following:
The talk will firstly provide a brief introduction into the main concepts of Reinforcement Learning, what are the main components and its limitations; it will then introduce sheeprl, our Deep Reinforcement Learning Framework, and its main features; finally it will showcase how to learn intelligent agents within the framework.
You can found us at:
Federico is a Data Scientist with extensive experience in deep learning and in particular in Computer Vision. Currently working in Orobix S.r.l., in the last two years he worked in the Reinforcement Learning field, where he led a small team during the development of sheeprl. He holds a M.Sc. from the University of Milano-Bicocca, where he worked in the Natural Language Processing field with a thesis on the semantic change of words during time.