معرفی شرکت ها


actorch-0.0.3


Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر

توضیحات

Deep reinforcement learning framework for fast prototyping based on PyTorch
ویژگی مقدار
سیستم عامل -
نام فایل actorch-0.0.3
نام actorch
نسخه کتابخانه 0.0.3
نگهدارنده []
ایمیل نگهدارنده []
نویسنده Luca Della Libera
ایمیل نویسنده luca.dellalib@gmail.com
آدرس صفحه اصلی https://github.com/lucadellalib/actorch
آدرس اینترنتی https://pypi.org/project/actorch/
مجوز Apache License 2.0
![logo](docs/_static/images/actorch-logo.png) [![Python version: 3.6 | 3.7 | 3.8 | 3.9 | 3.10](https://img.shields.io/badge/python-3.6%20|%203.7%20|%203.8%20|%203.9%20|%203.10-blue)](https://www.python.org/downloads/) [![Code style: black](https://img.shields.io/badge/code%20style-black-000000.svg)](https://github.com/psf/black) [![pre-commit](https://img.shields.io/badge/pre--commit-enabled-brightgreen?logo=pre-commit&logoColor=white)](https://github.com/pre-commit/pre-commit) Welcome to `actorch`, a deep reinforcement learning framework for fast prototyping based on [PyTorch](https://pytorch.org). The following algorithms are included: - [REINFORCE](https://people.cs.umass.edu/~barto/courses/cs687/williams92simple.pdf) - [Advantage Actor-Critic (A2C)](https://arxiv.org/abs/1602.01783) - [Actor-Critic Kronecker-Factored Trust Region (ACKTR)](https://arxiv.org/abs/1708.05144) - [Proximal Policy Optimization (PPO)](https://arxiv.org/abs/1707.06347) --------------------------------------------------------------------------------------------------------- ## 💡 Key features - Support for custom observation/action spaces - Support for custom multimodal input multimodal output (recurrent) models - Support for custom policy/value distributions - Support for custom preprocessing/postprocessing pipelines - Support for custom exploration strategies - Support for normalizing flows - Batched environments (both for training and evaluation) - Batched trajectory replay - Batched and distributional value estimation (e.g. distributional V-trace) - Data parallel and distributed data parallel multi-GPU training and evaluation - Automatic mixed precision training - Integration with [Ray Tune](https://docs.ray.io/en/latest/tune/index.html) for experiment execution and hyperparameter tuning at any scale - Effortless experiment definition through Python-based configuration files - Built-in visualizer to monitor experiment progress - Highly modular object-oriented design - Detailed API documentation --------------------------------------------------------------------------------------------------------- ## 🛠️️ Installation For Windows, make sure the latest [Visual C++ runtime](https://support.microsoft.com/en-us/help/2977003/the-latest-supported-visual-c-downloads) is installed. ### Using Pip First of all, install [Python](https://www.python.org). Open a terminal and run: ``` pip install actorch[visualizer] ``` If you don't need the `visualizer` (e.g. you are installing `actorch` on a headless server), run: ``` pip install actorch ``` ### Using Conda Clone or download and extract the repository, navigate to `<path-to-repository>/bin` and run the installation script (`install.sh` for Linux/macOS, `install.bat` for Windows). `actorch` (including the `visualizer`) and its dependencies (pinned to a specific version) will be installed in a [Conda](https://www.anaconda.com/) virtual environment named `actorch-env`. **NOTE**: you can directly use `actorch-env` and the `actorch` package in the local project directory for development (see [For development](#for-development)). ### Using Docker (Linux/macOS only) First of all, install [Docker](https://www.docker.com) and [NVIDIA Container Runtime](https://developer.nvidia.com/nvidia-container-runtime). Clone or download and extract the repository, navigate to `<path-to-repository>`, open a terminal and run: ``` docker build -t <desired-image-name> . # Build image docker run -it --runtime=nvidia <desired-image-name> # Run container from image ``` `actorch` (including the `visualizer`) and its dependencies (pinned to a specific version) will be installed in the specified Docker image. **NOTE**: you can directly use the `actorch` package in the local project directory inside a Docker container run from the specified Docker image for development (see [For development](#for-development)). ### From source First of all, install [Python](https://www.python.org). Clone or download and extract the repository, navigate to `<path-to-repository>`, open a terminal and run: ``` pip install .[visualizer] ``` If you don't need the `visualizer` (e.g. you are installing `actorch` on a headless server), run: ``` pip install . ``` ### For development First of all, install [Python](https://www.python.org) and [Git](https://git-scm.com/). Clone or download and extract the repository, navigate to `<path-to-repository>`, open a terminal and run: ``` pip install -e .[all] pre-commit install -f ``` This will install the package in editable mode (any change to the package in the local project directory will automatically reflect on the environment-wide package installed in the `site-packages` directory of your environment) along with its development, test and optional dependencies. Additionally, it installs a [git commit hook](https://git-scm.com/book/en/v2/Customizing-Git-Git-Hooks). Each time you commit, unit tests, static type checkers, code formatters and linters are run automatically. Run `pre-commit run --all-files` to check that the hook was successfully installed. For more details, see `pre-commit`'s [documentation](https://pre-commit.com). --------------------------------------------------------------------------------------------------------- ## ▶️ Quickstart In this example we will solve the [OpenAI Gym](https://www.gymlibrary.ml/) environment `CartPole-v1` using REINFORCE. Copy the following configuration in a file named `REINFORCE_CartPole-v1.py` (with the same indentation): ``` import gym from torch.optim import Adam from actorch import * experiment_params = ExperimentParams( run_or_experiment=REINFORCE, stop={"training_iteration": 30}, resources_per_trial={"cpu": 1, "gpu": 0}, checkpoint_freq=10, checkpoint_at_end=True, log_to_file=True, export_formats=["checkpoint", "model"], config=REINFORCE.Config( train_env_builder=lambda **config: ParallelBatchedEnv( lambda **config: gym.make("CartPole-v1", **config), config, num_workers=2, ), train_num_episodes_per_iteration=10, eval_interval_iterations=10, eval_env_config={"render_mode": None}, eval_num_episodes_per_iteration=10, policy_network_model_builder=FCNet, policy_network_model_config={ "torso_fc_configs": [ {"out_features": 64, "bias": True} ], }, policy_network_optimizer_builder=Adam, policy_network_optimizer_config={"lr": 1e-1}, discount=0.99, entropy_coeff=0.001, max_grad_l2_norm=0.5, seed=0, enable_amp=False, enable_reproducibility=True, log_sys_usage=True, suppress_warnings=False, ), ) ``` Open a terminal in the directory where you saved the configuration file and run: ``` pip install gym[classic_control] # Install dependencies for CartPole-v1 actorch run REINFORCE_CartPole-v1.py # Run experiment ``` Wait for a few minutes until the training has finished. The mean cumulative reward over the last 100 episodes should exceed 475, which means that the environment has been solved. You can now visualize the experiment progress stored in the generated [TensorBoard](https://www.tensorflow.org/tensorboard) files using [Plotly](https://plotly.com/): ``` cd experiments/REINFORCE_CartPole-v1/<auto-generated-experiment-name> actorch visualize plotly tensorboard ``` You can find the generated plots in `plots`. Congratulations, you have run your first experiment! **NOTE**: if you installed `actorch` in a virtual environment, you first need to activate it (`conda activate actorch-env` if you installed `actorch` using Conda). **HINT**: since a configuration file is a regular Python script, you can use all the features of the language (e.g. inheritance). --------------------------------------------------------------------------------------------------------- ## 📧 Contact [luca.dellalib@gmail.com](mailto:luca.dellalib@gmail.com) ---------------------------------------------------------------------------------------------------------


نیازمندی

مقدار نام
>=1.3.0 cloudpickle
>=0.4.3 colorama
>=1.4.0 GPUtil
>=0.25.0 gym
>=1.19.5 numpy
>=5.7.0 psutil
>=5.3.1 PyYAML
<2.*,>=1.13.0 ray[tune]
>=1.4.1 scipy
>=2.6.0 tensorboard
>=1.12.0 torch
>=3.8.3 flake8
>=2.6.0 pre-commit
>=0.812 mypy
>=3.2.0 pre-commit-hooks
>=3.2.1 matplotlib
>=4.9.0 plotly
>=22.3.0 black
>=2.3.1 cibuildwheel
>=3.3.0 twine
>=5.3.1 PyYAML
>=1.19.5 numpy
>=5.4.3 pytest
>=5.4.2 isort
>=4.48.2 tqdm
>=2.9.0 pytest-cov
>=2.6.0 tensorboard
>=1.4.1 scipy
>=20.1.4 flake8-bugbear
>=22.3.0 black
>=2.3.1 cibuildwheel
>=3.8.3 flake8
>=20.1.4 flake8-bugbear
>=5.4.2 isort
>=0.812 mypy
>=2.6.0 pre-commit
>=3.2.0 pre-commit-hooks
>=3.3.0 twine
>=5.4.3 pytest
>=2.9.0 pytest-cov
>=3.2.1 matplotlib
>=1.19.5 numpy
>=4.9.0 plotly
>=5.3.1 PyYAML
>=1.4.1 scipy
>=2.6.0 tensorboard
>=4.48.2 tqdm


زبان مورد نیاز

مقدار نام
>=3.6 Python


نحوه نصب


نصب پکیج whl actorch-0.0.3:

    pip install actorch-0.0.3.whl


نصب پکیج tar.gz actorch-0.0.3:

    pip install actorch-0.0.3.tar.gz