Project Name | Stars | Downloads | Repos Using This | Packages Using This | Most Recent Commit | Total Releases | Latest Release | Open Issues | License | Language |
---|---|---|---|---|---|---|---|---|---|---|
Keras | 58,563 | 330 | 9 hours ago | 68 | May 13, 2022 | 391 | apache-2.0 | Python | ||
Deep Learning for humans | ||||||||||
Data Science Ipython Notebooks | 25,025 | a month ago | 33 | other | Python | |||||
Data science Python notebooks: Deep learning (TensorFlow, Theano, Caffe, Keras), scikit-learn, Kaggle, big data (Spark, Hadoop MapReduce, HDFS), matplotlib, pandas, NumPy, SciPy, Python essentials, AWS, and various command lines. | ||||||||||
Netron | 23,121 | 4 | 63 | 20 hours ago | 489 | July 04, 2022 | 28 | mit | JavaScript | |
Visualizer for neural network, deep learning, and machine learning models | ||||||||||
Mask_rcnn | 23,049 | a month ago | 5 | March 05, 2019 | 1,947 | other | Python | |||
Mask R-CNN for object detection and instance segmentation on Keras and TensorFlow | ||||||||||
100 Days Of Ml Code | 19,753 | a year ago | 9 | mit | Jupyter Notebook | |||||
100-Days-Of-ML-Code中文版 | ||||||||||
D2l En | 18,092 | 2 days ago | 106 | other | Python | |||||
Interactive deep learning book with multi-framework code, math, and discussions. Adopted at 400 universities from 60 countries including Stanford, MIT, Harvard, and Cambridge. | ||||||||||
Ncnn | 17,307 | 2 days ago | 19 | July 01, 2022 | 989 | other | C++ | |||
ncnn is a high-performance neural network inference framework optimized for the mobile platform | ||||||||||
Onnx | 14,893 | 148 | 245 | 13 hours ago | 26 | June 18, 2022 | 388 | apache-2.0 | Python | |
Open standard for machine learning interoperability | ||||||||||
Best Of Ml Python | 13,788 | 2 days ago | 19 | cc-by-sa-4.0 | ||||||
🏆 A ranked list of awesome machine learning Python libraries. Updated weekly. | ||||||||||
Horovod | 13,342 | 20 | 10 | 3 days ago | 72 | June 21, 2022 | 347 | other | Python | |
Distributed training framework for TensorFlow, Keras, PyTorch, and Apache MXNet. |
RL-botics is a toolbox with highly optimized implementations of Deep Reinforcement Learning algorithms for robotics developed with Keras and TensorFlow in Python3.
The objective was to have modular, clean and easy to read codebase so that the research community may build on top with ease. The implementations can be integrated with OpenAI Gym environments. The majority of the algorithms are Policy Search Methods as the toolbox is targetted for robotic applications.
Requirements:
It is highly recommended to install this package in a virtual environment, such as Miniconda. Please find the Conda installation here.
To create a new conda environment called RL
:
conda create -n RL python=3
To activate the environment:
source activate RL
To deactivate the environment:
source deactivate
To install the package, we recommend cloning the original package:
git clone https://github.com/Suman7495/rl-botics.git
cd rl-botics
pip install -e .
To run any algorithm in the default setting, simply run:
cd rl_botics/<algo>/
python main.py
For example, to run TRPO:
cd rl_botics/trpo/
python main.py
Numerous other options can be added too, but it is recommended to modify the hyerperparameters in hyperparameters.py
.
The algorithms implemented are:
To be added:
All environments are in the envs
directory. The environments available currently are:
All the algorithms are in the rl_botics
directory. Each algorithm specified above has an individual directory.
The directory common
contains common modular classes to easily build new algorithms.
approximators
: Basic Deep Neural Networks (Dense, Conv, LSTM).data_collection
: Performs rollouts and collect observations and rewardslogger
: Log training data and other informationplotter
: Plot graphspolicies
: Common policies such as Random, Softmax, Parametrized Softmax and Gaussian Policyutils
: Functions to compute the expected return, the Generalized Advantage Estimation (GAE), etc.Each algorithm directory contains at least 3 files:
main.py
: Main script to run the algorithmhyperparameters.py
: File to contain the default hyperparameters<algo>.py
: Implementation of the algorithmutils.py
: (Optional) File containing some utility functionsSome algorithm directories may have additional files specific to the algorithm.
To contribute to this package, it is recommended to follow this structure:
main.py
should contain at least the following functions:
main
: Parses input argument, builds the environment and agent, and train the agent.argparse
: Parses input argument and loads default hyperparameters from hyperparameter.py
.<algo>.py
should contain at least the following methods:
__init__
: Initializes the classes_build_graph
: Calls the following methods to build the TensorFlow graph:
_init_placeholders
: Initialize TensorFlow placeholders_build_policy
: Build policy TensorFlow graph_build_value_function
: Build value function TensorFlow graph_loss
: Build policy loss function TensorFlwo graphtrain
: Main training loop called by main.py
update_policy
: Update the policyupdate_value
: Update the value functionprint_results
: Print the training resultsprocess_paths
: (optional) Process collected trajectories to return the feed dictionary for TensorFlowIt is recommended to check the structure of ppo.py
and follow a similar structure.
Suman Pal
MIT License.