|
|
.PHONY: style check_code_quality |
|
|
|
|
|
export PYTHONPATH = . |
|
|
check_dirs := inference inference_sdk |
|
|
|
|
|
style: |
|
|
black $(check_dirs) |
|
|
isort --profile black $(check_dirs) |
|
|
|
|
|
check_code_quality: |
|
|
black --check $(check_dirs) |
|
|
isort --check-only --profile black $(check_dirs) |
|
|
|
|
|
flake8 $(check_dirs) --count --select=E9,F63,F7,F82 --show-source --statistics |
|
|
|
|
|
flake8 $(check_dirs) --count --max-line-length=88 --exit-zero --ignore=D --extend-ignore=E203,E501,W503 --statistics |
|
|
|
|
|
start_test_docker_cpu: |
|
|
docker run -d --rm -p $(PORT):$(PORT) -e PORT=$(PORT) -e MAX_BATCH_SIZE=17 --name inference-test roboflow/${INFERENCE_SERVER_REPO}:test |
|
|
|
|
|
start_test_docker_gpu: |
|
|
docker run -d --rm -p $(PORT):$(PORT) -e PORT=$(PORT) -e MAX_BATCH_SIZE=17 --gpus=all --name inference-test roboflow/${INFERENCE_SERVER_REPO}:test |
|
|
|
|
|
start_test_docker_jetson: |
|
|
docker run -d --rm -p $(PORT):$(PORT) -e PORT=$(PORT) -e MAX_ACTIVE_MODELS=1 -e MAX_BATCH_SIZE=17 --runtime=nvidia --name inference-test roboflow/${INFERENCE_SERVER_REPO}:test |
|
|
|
|
|
stop_test_docker: |
|
|
docker rm -f inference-test |
|
|
|
|
|
create_wheels: |
|
|
python -m pip install --upgrade pip |
|
|
python -m pip install wheel twine requests -r requirements/_requirements.txt |
|
|
rm -f dist/* |
|
|
python .release/pypi/inference.core.setup.py bdist_wheel |
|
|
python .release/pypi/inference.cpu.setup.py bdist_wheel |
|
|
python .release/pypi/inference.gpu.setup.py bdist_wheel |
|
|
python .release/pypi/inference.setup.py bdist_wheel |
|
|
python .release/pypi/inference.sdk.setup.py bdist_wheel |
|
|
python .release/pypi/inference.cli.setup.py bdist_wheel |
|
|
|
|
|
upload_wheels: |
|
|
twine upload dist/*.whl |
|
|
|