معرفی شرکت ها


bentoctl-0.3.4


Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر
Card image cap
تبلیغات ما

مشتریان به طور فزاینده ای آنلاین هستند. تبلیغات می تواند به آنها کمک کند تا کسب و کار شما را پیدا کنند.

مشاهده بیشتر

توضیحات

Fast model deployment with BentoML on cloud platforms.
ویژگی مقدار
سیستم عامل -
نام فایل bentoctl-0.3.4
نام bentoctl
نسخه کتابخانه 0.3.4
نگهدارنده []
ایمیل نگهدارنده []
نویسنده bentoml.org
ایمیل نویسنده contact@bentoml.com
آدرس صفحه اصلی https://github.com/bentoml/bentoctl
آدرس اینترنتی https://pypi.org/project/bentoctl/
مجوز Apache-2.0
<div align="center"> <h1>bentoctl</h1> <i>Fast model deployment with BentoML on cloud platforms</i> <p> <img alt="PyPI" src="https://img.shields.io/pypi/v/bentoctl?style=flat-square"> <img alt="GitHub branch checks state" src="https://img.shields.io/github/checks-status/bentoml/bentoctl/main?style=flat-square"> <img alt="Codecov" src="https://img.shields.io/codecov/c/github/bentoml/bentoctl?style=flat-square"> </p> </div> <br> `bentoctl` is a CLI tool for deploying your machine-learning models to any cloud platforms. It built on top of [BentoML: the unified model serving framework](https://github.com/bentoml/BentoML), and makes it easy to bring any BentoML packaged model to production. 👉 [Pop into our Slack community!](https://l.linklyhq.com/l/ktPp) We're happy to help with any issue you face or even just to meet you and hear what you're working on :) ## Features: * Supports major cloud providers: AWS, Azure, Google Cloud, and more. * Easy to deploy, update and reproduce model deployments. * First class integration with Terraform. * Optimized for CI/CD workflow. * Extensible with custom operators. * High performance serving powered by [BentoML](https://github.com/bentoml/BentoML) ## Supported Platforms: * [AWS Lambda](https://github.com/bentoml/aws-lambda-deploy) * [AWS SageMaker](https://github.com/bentoml/aws-sagemaker-deploy) * [AWS EC2](https://github.com/bentoml/aws-ec2-deploy) * [Google Cloud Run](https://github.com/bentoml/google-cloud-run-deploy) * [Google Compute Engine](https://github.com/bentoml/google-compute-engine-deploy) * [Azure Container Instances](https://github.com/bentoml/azure-container-instances-deploy) * [Heroku](https://github.com/bentoml/heroku-deploy) * Looking for **Kubernetes**? Try out [Yatai: Model deployment at scale on Kubernetes](https://github.com/bentoml/Yatai). * **Customize deploy target** by creating bentoctl plugin from the [deployment operator template](https://github.com/bentoml/bentoctl-operator-template). **Upcoming:** * [Azure Functions](https://github.com/bentoml/azure-functions-deploy) * [Knative](https://github.com/bentoml/bentoctl/issues/79) ## Install bentoctl ```bash pip install bentoctl ``` | 💡 bentoctl designed to work with BentoML version 1.0.0 and above. For BentoML 0.13 or below, you can use the `pre-v1.0` branch in the operator repositories and follow the instruction in the README. You can also check out the quickstart guide for 0.13 [here](./docs/013-deployment.md). ## Next steps - [Quickstart Guide](./docs/quickstart.md) walks through a series of steps to deploy a bento to AWS Lambda as API server. - [Core Concepts](./docs/core-concepts.md) explains the core concepts in bentoctl. - [Operator List](./docs/operator-list.md) lists official operators and their current status. ## Community - To report a bug or suggest a feature request, use [GitHub Issues](https://github.com/bentoml/bentoctl/issues/new/choose). - For other discussions, use [Github Discussions](https://github.com/bentoml/BentoML/discussions) under the [BentoML repo](https://github.com/bentoml/BentoML/) - To receive release announcements and get support, join us on [Slack](http://join.slack.bentoml.org). ## Contributing There are many ways to contribute to the project: - Create and share new operators. Use [deployment operator template](https://github.com/bentoml/bentoctl-operator-template) to get started. - If you have any feedback on the project, share it with the community in [Github Discussions](https://github.com/bentoml/BentoML/discussions) under the [BentoML repo](https://github.com/bentoml/BentoML/). - Report issues you're facing and "Thumbs up" on issues and feature requests that are relevant to you. - Investigate bugs and reviewing other developer's pull requests. ## Usage Reporting BentoML and bentoctl collects usage data that helps our team to improve the product. Only bentoctl's CLI commands calls are being reported. We strip out as much potentially sensitive information as possible, and we will never collect user code, model data, model names, or stack traces. Here's the [code](./bentoctl/utils/usage_stats.py) for usage tracking. You can opt-out of usage tracking by setting environment variable `BENTOML_DO_NOT_TRACK=True`: ```bash export BENTOML_DO_NOT_TRACK=True ```


نیازمندی

مقدار نام
>=1.0.0,<2.0.0 bentoml
>=8,<9 click
>=1,<2 cerberus
- rich
>=6,<7 PyYaml
==0.4.4 simple-term-menu
ython_versio importlib-metadata;
>=5,<6 docker
>=2.9.0,<3.0.0 semantic_version


زبان مورد نیاز

مقدار نام
>=3.7,<4.0 Python


نحوه نصب


نصب پکیج whl bentoctl-0.3.4:

    pip install bentoctl-0.3.4.whl


نصب پکیج tar.gz bentoctl-0.3.4:

    pip install bentoctl-0.3.4.tar.gz