Skip to content

Commit

Permalink
remake api html
Browse files Browse the repository at this point in the history
  • Loading branch information
hsywhu committed Sep 17, 2019
1 parent 3d4ce8a commit cab457b
Show file tree
Hide file tree
Showing 6 changed files with 386 additions and 63 deletions.
12 changes: 6 additions & 6 deletions docs/rst/README (1).rst
Original file line number Diff line number Diff line change
Expand Up @@ -19,16 +19,16 @@ conveniently into RLCard with the above design principles.
User Guide
==========

- `Toy examples <docs/toy-examples.md>`__
- `RLCard high-level design <docs/high-level-design.md>`__
- `Games in RLCard <docs/games.md>`__
- `Algorithms in RLCard <docs/algorithms.md>`__
- `Developping new algorithms <docs/developping-algorithms.md>`__
- `Toy examples <toy-examples.md>`__
- `RLCard high-level design <high-level-design.md>`__
- `Games in RLCard <games.md>`__
- `Algorithms in RLCard <algorithms.md>`__
- `Developping new algorithms <developping-algorithms.md>`__

Developer Guide
===============

- `Adding new environments <docs/adding-new-environments.md>`__
- `Adding new environments <adding-new-environments.md>`__

Application Programming Interface (API)
=======================================
Expand Down
49 changes: 36 additions & 13 deletions docs/rst/README.rst
Original file line number Diff line number Diff line change
@@ -1,28 +1,37 @@
RLCard: A Toolkit for Reinforcement Learning in Card Games
==========================================================

RLCard is an opensource toolkit for developing Reinforcement Learning
|Build Status| |Codacy Badge| |Coverage Status|

RLCard is a opensource toolkit for developing Reinforcement Learning
(RL) algorithms in card games. It supports multiple challenging card
game environments with common and easy-to-use interfaces. The goal of
the toolkit is to enable more people to study game AI and push forward
the research of imperfect information games. RLCard is developed by
`DATA Lab <http://faculty.cs.tamu.edu/xiahu/>`__ at Texas A&M
University.
University. **NOTE: The project is still in final testing!**

Installation
============
------------

Make sure that you have **Python 3.5+** and **pip** installed. You can
install ``rlcard`` with ``pip`` as follow:

::
.. code:: console
git clone https://github.com/datamllab/rlcard.git
cd rlcard
pip install -e .
To check whether it is intalled correctly, try the example with random
agents:

.. code:: console
python examples/blackjack_random.py
Getting Started
===============
---------------

The interfaces generally follow `OpenAI
gym <https://github.com/openai/gym>`__ style. We recommend starting with
Expand All @@ -35,14 +44,14 @@ Blackjack <docs/toy-examples.md#deep-q-learning-on-blackjack>`__ \*
For more examples, please refer to `examples/ <examples>`__.

Documents
=========
---------

Please refer to the `Documents <docs/README.md>`__ for general concepts
introduction. API documents are available at our `github
page <https://rlcard.github.io/index.html>`__.

Available Environments
======================
----------------------

The table below shows the environments that are (or will be soon)
available in RLCard. We provide a complexity estimation for the games on
Expand All @@ -68,19 +77,33 @@ game environment.
+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------------------+---------------------+---------------+-------------------+-------------+
| UNO (`wiki <https://en.wikipedia.org/wiki/Uno_(card_game>`__, `baike <https://baike.baidu.com/item/UNO%E7%89%8C/2249587>`__) | 10^163 | 10^10 | 10^1 | - | Come soon |
+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------------------+---------------------+---------------+-------------------+-------------+
| Sheng Ji (`wiki <https://en.wikipedia.org/wiki/Sheng_ji>`__, `baike <https://baike.baidu.com/item/%E5%8D%87%E7%BA%A7/3563150>`__) | 10^173 ~ 10^180 | 10^61 | 10^13 | - | Come soon |
| Sheng Ji (`wiki <https://en.wikipedia.org/wiki/Sheng_ji>`__, `baike <https://baike.baidu.com/item/%E5%8D%87%E7%BA%A7/3563150>`__) | 10^157 ~ 10^165 | 10^61 | 10^13 | - | Come soon |
+--------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------+-------------------+---------------------+---------------+-------------------+-------------+

Evaluation
----------

We wrap a ``Logger`` that conveniently saves/plots the results. Example
outputs are as follows: |Learning Curves|

Disclaimer
==========
----------

Please note that this is a **pre-release** version of the RLCard. The
toolkit is provided "**as is**," without warranty of any kind, express
or implied, including but not limited to the warranties of
merchantability, fitness for a particular purpose and noninfringement.

Acknowledgements
================

We would like to thank JJ World Network Technology Co.,LTD for the
support.
----------------

We would like to thank JJ World Network Technology Co.,LTD for technical
the support.

.. |Build Status| image:: https://travis-ci.org/datamllab/RLCard.svg?branch=master
:target: https://travis-ci.org/datamllab/RLCard
.. |Codacy Badge| image:: https://api.codacy.com/project/badge/Grade/248eb15c086748a4bcc830755f1bd798
:target: https://www.codacy.com/manual/daochenzha/rlcard?utm_source=github.com&utm_medium=referral&utm_content=datamllab/rlcard&utm_campaign=Badge_Grade
.. |Coverage Status| image:: https://coveralls.io/repos/github/datamllab/rlcard/badge.svg?branch=master
:target: https://coveralls.io/github/datamllab/rlcard?branch=master
.. |Learning Curves| image:: docs/imgs/curves.png
13 changes: 11 additions & 2 deletions docs/rst/algorithms.rst
Original file line number Diff line number Diff line change
Expand Up @@ -13,6 +13,15 @@ Index
feeding the state into the model.
- ``Memory``: A memory buffer that manages the storing and sampling of
transitions.
- ``Estimator``: The neural network that is used to make predictions. #
DeepCFR test
- ``Estimator``: The neural network that is used to make predictions.

DeepCFR
=======

Deep Counterfactual Regret Minimization (DeepCFR) is a state-of-the-art
framework for solving imperfect-information games. We wrap DeepCFR as an
example to show how state-of-the-art framework can be connected to the
environments. In the DeepCFR, the following classes are implemented: \*
``DeepCFR``: The DeepCFR class that interacts with the environment. \*
``Fixed Size Ring Buffer``: A memory buffer that manages the storing and
sampling of transitions.
Loading

0 comments on commit cab457b

Please sign in to comment.