.PHONY: style check_code_quality export PYTHONPATH = . check_dirs := inference inference_sdk style: black $(check_dirs) isort --profile black $(check_dirs) check_code_quality: black --check $(check_dirs) isort --check-only --profile black $(check_dirs) # stop the build if there are Python syntax errors or undefined names flake8 $(check_dirs) --count --select=E9,F63,F7,F82 --show-source --statistics # exit-zero treats all errors as warnings. E203 for black, E501 for docstring, W503 for line breaks before logical operators flake8 $(check_dirs) --count --max-line-length=88 --exit-zero --ignore=D --extend-ignore=E203,E501,W503 --statistics start_test_docker_cpu: docker run -d --rm -p $(PORT):$(PORT) -e PORT=$(PORT) -e MAX_BATCH_SIZE=17 --name inference-test roboflow/${INFERENCE_SERVER_REPO}:test start_test_docker_gpu: docker run -d --rm -p $(PORT):$(PORT) -e PORT=$(PORT) -e MAX_BATCH_SIZE=17 --gpus=all --name inference-test roboflow/${INFERENCE_SERVER_REPO}:test start_test_docker_jetson: docker run -d --rm -p $(PORT):$(PORT) -e PORT=$(PORT) -e MAX_ACTIVE_MODELS=1 -e MAX_BATCH_SIZE=17 --runtime=nvidia --name inference-test roboflow/${INFERENCE_SERVER_REPO}:test stop_test_docker: docker rm -f inference-test create_wheels: python -m pip install --upgrade pip python -m pip install wheel twine requests -r requirements/_requirements.txt rm -f dist/* python .release/pypi/inference.core.setup.py bdist_wheel python .release/pypi/inference.cpu.setup.py bdist_wheel python .release/pypi/inference.gpu.setup.py bdist_wheel python .release/pypi/inference.setup.py bdist_wheel python .release/pypi/inference.sdk.setup.py bdist_wheel python .release/pypi/inference.cli.setup.py bdist_wheel upload_wheels: twine upload dist/*.whl