#
alphago
Here are 76 public repositories matching this topic...
evg-tyurin
commented
Feb 10, 2019
During self-play phase we usually collect different examples for the same board states. Should we preprocess such examples before optimizing the NNet? In the current implementation, we don't preprocess them so we train NNet and expect different output from the same input values. I think this may be wrong.
PyTorch implementation of DQN, AC, ACER, A2C, A3C, PG, DDPG, TRPO, PPO, SAC, TD3 and ....
algorithm
deep-learning
deep-reinforcement-learning
pytorch
dqn
policy-gradient
sarsa
resnet
a3c
reinforce
sac
alphago
actor-critic
trpo
ppo
a2c
actor-critic-algorithm
td3
-
Updated
Mar 18, 2020 - Python
BetaGo: AlphaGo for the masses, live on GitHub.
-
Updated
Nov 18, 2019 - Python
Code and other material for the book "Deep Learning and the Game of Go"
python
data-science
machine-learning
games
deep-learning
neural-networks
game-of-go
alphago
alphago-zero
-
Updated
Apr 9, 2020 - Python
A Go A.I. based on MCTS(AlphaGo and DeepZenGo's basic algorithm) WITHOUT Deep Learning
-
Updated
Jan 15, 2019 - C++
A Go playing program implemented in Tensorflow roughly according to the architecture of AlphaGo. Current strength is 3~4 amateur dan.
-
Updated
Mar 18, 2019 - Python
A student implementation of Alpha Go Zero
-
Updated
Aug 1, 2018 - Python
中国象棋alpha zero程序
-
Updated
Jan 2, 2019 - Jupyter Notebook
datasets for computer go
go
sgf
alphago
computer-go
tygem
computer-go-dataset
fineart
alphazero
minigo
phoenixgo
leelazero
muzero
golaxy
elf-opengo
-
Updated
May 13, 2020 - C++
An asynchronous/parallel method of AlphaGo Zero algorithm with Gomoku
tree
algorithm
board
tensorflow
gpu
paper
parallel
deep-reinforcement-learning
mcts
gomoku
noise
tree-search
tensorlayer
alphago
mpi4py
dirichlet-distribution
alphazero
alphazero-gomoku
board-model
playout-times
add-noises
junxiaosong
-
Updated
Jan 20, 2020 - Python
A Deep Learning UCI-Chess Variant Engine written in C++ & Python 🐦
python
open-source
machine-learning
chess-engine
deep-learning
mxnet
artificial-intelligence
mcts
gluon
lichess
convolutional-neural-network
alphago
python-chess
alphazero
crazyhouse
-
Updated
May 28, 2020 - Jupyter Notebook
A minimalist deep learning library in Javascript using WebGL + asm.js. Run convolutional neural network in your browser.
-
Updated
Nov 21, 2017 - JavaScript
AlphaGo Zero implementation using Flux.jl
-
Updated
Jan 16, 2019 - Julia
Reinforcing Your Learning of Reinforcement Learning
reinforcement-learning
tic-tac-toe
space-invaders
q-learning
doom
dqn
mcts
policy-gradient
cartpole
gomoku
ddpg
atari-2600
alphago
frozenlake
ppo
advantage-actor-critic
alphago-zero
-
Updated
Jul 14, 2019 - Python
docker
kubernetes
board-game
flash
deep-neural-networks
ai
microservice
game-engine
wiki
actionscript
deep-reinforcement-learning
cnn
dnn
mcts
finite-state-machine
deeplearning
starling
fuzzy-logic-control
alphago
policytree
-
Updated
Nov 22, 2019 - HTML
-
Updated
Jan 12, 2019 - Python
alphaGo版本的五子棋(gobang, gomoku)
-
Updated
Mar 17, 2020 - Python
9x9 AlphaGo
-
Updated
Jul 27, 2016 - Python
Othello AI (AlphaGo's PV-MCTS algorithm)
-
Updated
Nov 7, 2018 - Python
c++ implementation of alphagozero
-
Updated
May 29, 2018 - C++
-
Updated
May 17, 2020 - C++
CrazyAra - A Deep Learning UCI-Chess Variant Engine written in C++ 🐦
open-source
machine-learning
chess-engine
deep-learning
mxnet
cpp
artificial-intelligence
mcts
lichess
convolutional-neural-network
alphago
alphazero
crazyhouse
chess-variants
-
Updated
Oct 8, 2019 - C++
This is a tutorial written for Caffe2 which mocks google AlphaGo Fan and AlphaGo Zero.
-
Updated
Dec 20, 2018 - Jupyter Notebook
Tic Tac Toe with Alpha Zero method - My first work
-
Updated
Aug 23, 2018 - Python
for the momery of ZheLi
go
docker
kubernetes
golang
machine-learning
microservices
deep-neural-networks
ai
deep-learning
microservice
docker-compose
fabric
tensorflow
rxjava
blockchain
fintech
hyperledger
restfull-api
darknet
alphago
-
Updated
Apr 4, 2018 - JavaScript
Improve this page
Add a description, image, and links to the alphago topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with the alphago topic, visit your repo's landing page and select "manage topics."
首先感谢分享程序。
请问在6x6 四子棋的训练过程中,有没有调节learning_rate或者其他参数?
程序里的c_puct=5 , 温度t=1,学习率 0.002,batch_size 512 , deque最大长度10000, kl-targ=0.02 ,epochs=5
我使用你程序里的预设参数 tensorflow训练6x6 四子棋 ,loss下降到2左右就无法下降了,调节学习率也没成功。。。求帮助解答,谢谢
另外,不明白explain_var_old这个参考数值的意义。