TY - GEN
T1 - Multi-Objective Optimization Based on AlphaZero Method Applied to Connection Games
AU - Chiu, Hsuan Kai
AU - Chen, Chih Hung
AU - Lin, Shun Shii
N1 - Publisher Copyright:
© 2024 IEEE.
PY - 2024
Y1 - 2024
N2 - Open-ended learning, introduced by the DeepMind team in 2021, represents a novel AI approach to achieve multi-objective optimization, allowing it to simultaneously handle multiple tasks, whereas traditional AI is optimized only for a single task. This study describes an AI implementation similar to open-ended learning using relatively familiar technologies and game rules, making it adaptable to two different game rule sets. Experimental results indicate that, given equal training time, multi-objective AlphaZero outperforms the single-objective optimized version of AlphaZero. Training using both rule sets impacts the neural network. Our results indicate that simpler training data can accelerate the model's initial learning phase without compromising performance across the two rule sets. Multi-objective training is found to enhance the overall efficiency and effectiveness of neural network learning. Specifically, incorporating simpler rule sets accelerates early-stage training and helps lay the foundation for learning more complex rules. The findings suggest that leveraging the interaction between different levels of rule complexity can achieve more balanced and comprehensive training outcomes, ultimately resulting in a more generalizable AI model.
AB - Open-ended learning, introduced by the DeepMind team in 2021, represents a novel AI approach to achieve multi-objective optimization, allowing it to simultaneously handle multiple tasks, whereas traditional AI is optimized only for a single task. This study describes an AI implementation similar to open-ended learning using relatively familiar technologies and game rules, making it adaptable to two different game rule sets. Experimental results indicate that, given equal training time, multi-objective AlphaZero outperforms the single-objective optimized version of AlphaZero. Training using both rule sets impacts the neural network. Our results indicate that simpler training data can accelerate the model's initial learning phase without compromising performance across the two rule sets. Multi-objective training is found to enhance the overall efficiency and effectiveness of neural network learning. Specifically, incorporating simpler rule sets accelerates early-stage training and helps lay the foundation for learning more complex rules. The findings suggest that leveraging the interaction between different levels of rule complexity can achieve more balanced and comprehensive training outcomes, ultimately resulting in a more generalizable AI model.
KW - AlphaZero
KW - Five-in-a-row
KW - Four-in-a-row
KW - Multi-objective optimization
KW - Open-ended learning
UR - http://www.scopus.com/inward/record.url?scp=85210246540&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=85210246540&partnerID=8YFLogxK
U2 - 10.1109/AIC61668.2024.10731063
DO - 10.1109/AIC61668.2024.10731063
M3 - Conference contribution
AN - SCOPUS:85210246540
T3 - 2024 IEEE 3rd World Conference on Applied Intelligence and Computing, AIC 2024
SP - 1249
EP - 1254
BT - 2024 IEEE 3rd World Conference on Applied Intelligence and Computing, AIC 2024
PB - Institute of Electrical and Electronics Engineers Inc.
T2 - 3rd IEEE World Conference on Applied Intelligence and Computing, AIC 2024
Y2 - 27 June 2024 through 28 June 2024
ER -