RoboCup soccer competitions are considered among the most challenging multi-robot adversarial environments, due to their high dynamism and the partial observability of the environment. In this paper we introduce a method based on a combination of Monte Carlo search and data aggregation (MCSDA) to adapt discrete-action soccer policies for a defender robot to the strategy of the opponent team. By exploiting a simple representation of the domain, a supervised learning algorithm is trained over an initial collection of data consisting of several simulations of human expert policies. Monte Carlo policy rollouts are then generated and aggregated to previous data to improve the learned policy over multiple epochs and games. The proposed approach has been extensively tested both on a soccer-dedicated simulator and on real robots. Using this method, our learning robot soccer team achieves an improvement in ball interceptions, as well as a reduction in the number of opponents’ goals. Together with a better performance, an overall more efficient positioning of the whole team within the field is achieved.

Using Monte Carlo Search With Data Aggregation to Improve Robot Soccer Policies / Riccio, Francesco; Capobianco, Roberto; Nardi, Daniele. - 9776:(2017), pp. 256-267. (Intervento presentato al convegno 20th Annual RoboCup International Symposium, 2016 tenutosi a Leipzig; Germany) [10.1007/978-3-319-68792-6_21].

Using Monte Carlo Search With Data Aggregation to Improve Robot Soccer Policies

RICCIO, FRANCESCO
;
CAPOBIANCO, ROBERTO
;
NARDI, Daniele
2017

Abstract

RoboCup soccer competitions are considered among the most challenging multi-robot adversarial environments, due to their high dynamism and the partial observability of the environment. In this paper we introduce a method based on a combination of Monte Carlo search and data aggregation (MCSDA) to adapt discrete-action soccer policies for a defender robot to the strategy of the opponent team. By exploiting a simple representation of the domain, a supervised learning algorithm is trained over an initial collection of data consisting of several simulations of human expert policies. Monte Carlo policy rollouts are then generated and aggregated to previous data to improve the learned policy over multiple epochs and games. The proposed approach has been extensively tested both on a soccer-dedicated simulator and on real robots. Using this method, our learning robot soccer team achieves an improvement in ball interceptions, as well as a reduction in the number of opponents’ goals. Together with a better performance, an overall more efficient positioning of the whole team within the field is achieved.
2017
20th Annual RoboCup International Symposium, 2016
Humanoid robots; Multi-robot systems;Policy learning; Reinforcement learning
04 Pubblicazione in atti di convegno::04b Atto di convegno in volume
Using Monte Carlo Search With Data Aggregation to Improve Robot Soccer Policies / Riccio, Francesco; Capobianco, Roberto; Nardi, Daniele. - 9776:(2017), pp. 256-267. (Intervento presentato al convegno 20th Annual RoboCup International Symposium, 2016 tenutosi a Leipzig; Germany) [10.1007/978-3-319-68792-6_21].
File allegati a questo prodotto
File Dimensione Formato  
Riccio_Using-Monte-Carlo_2017.pdf

solo gestori archivio

Tipologia: Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza: Tutti i diritti riservati (All rights reserved)
Dimensione 1.16 MB
Formato Adobe PDF
1.16 MB Adobe PDF   Contatta l'autore

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/11573/928050
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 2
  • ???jsp.display-item.citation.isi??? ND
social impact