Evolving neural networks to play checkers without relying on expert knowledge

被引:96
作者
Chellapilla, K [1 ]
Fogel, DB
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, La Jolla, CA 92093 USA
[2] Nat Select Inc, La Jolla, CA 92037 USA
来源
IEEE TRANSACTIONS ON NEURAL NETWORKS | 1999年 / 10卷 / 06期
关键词
alpha-beta search; checkers; evolutionary computation; feedforward neural networks; game playing;
D O I
10.1109/72.809083
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
An experiment was conducted where neural networks compete for survival in an evolving population based on their ability to play checkers. More specifically, multilayer feedforward neural networks were used to evaluate alternative board positions and games were played using a minimax search strategy. At each generation, the extant neural networks were paired in competitions and selection was used to eliminate those that performed poorly relative to other networks. Offspring neural networks were created from the survivors using random variation of all weights and bias terms. After a series of 250 generations, the best-evolved neural network was played against human opponents in a series of 90 games on an internet website, The neural network was able to defeat two expert-level players and played to a draw against a master. The final rating of the neural network placed it in the "Class A" category using a standard rating system. Of particular importance in the design of the experiment was the fact that no features beyond the piece differential were given to the neural networks as a priori knowledge. The process of evolution was able to extract all of the additional information required to play at this level of competency. It accomplished this based almost solely on the feedback offered in the final aggregated outcome of each game played (i.e., win, lose, or draw). This procedure stands in marked contrast to the typical artifice of explicitly injecting expert knowledge into a game-playing program.
引用
收藏
页码:1382 / 1391
页数:10
相关论文
共 13 条
[1]  
Clark D, 1997, IEEE EXPERT, V12, P31
[2]  
Fogel D.B., 1995, EVOLUTIONARY COMPUTA
[3]  
GRIFFITH NJL, 1997, UNPUB NEURODRAUGHTS
[4]  
Kaindl H., 1990, COMPUTERS CHESS COGN, P133, DOI [10.1007/978-1-4613-9080-0, DOI 10.1007/978-1-4613-9080-0]
[5]   STEPS TOWARD ARTIFICIAL INTELLIGENCE [J].
MINSKY, M .
PROCEEDINGS OF THE INSTITUTE OF RADIO ENGINEERS, 1961, 49 (01) :8-+
[6]   CELL SIGNALING PATHWAY INVOLVED IN PACAP-INDUCED AR4-2J CELL-PROLIFERATION [J].
MORISSET, J ;
DOUZIECH, N ;
RYDZEWSKA, G ;
BUSCAIL, L ;
RIVARD, N .
CELLULAR SIGNALLING, 1995, 7 (03) :195-205
[7]   Co-evolution in the successful learning of backgammon strategy [J].
Pollack, JB ;
Blair, AD .
MACHINE LEARNING, 1998, 32 (03) :225-240
[8]  
SAMUEL AL, 1959, IBM J RES DEV, V3, P210
[9]   A WORLD CHAMPIONSHIP CALIBER CHECKERS PROGRAM [J].
SCHAEFFER, J ;
CULBERSON, J ;
TRELOAR, N ;
KNIGHT, B ;
LU, P ;
SZAFRON, D .
ARTIFICIAL INTELLIGENCE, 1992, 53 (2-3) :273-289
[10]  
Schaeffer J., 1996, AI Magazine, V17, P21