wyzely-detect/pyproject.toml

94 lines
3.3 KiB
TOML
Raw Normal View History

[tool.poetry]
name = "wyzely-detect"
version = "0.2.1"
description = "Recognize faces/objects in a video stream (from a webcam or a security camera) and send notifications to your devices"
authors = ["slashtechno <77907286+slashtechno@users.noreply.github.com>"]
repository = "https://github.com/slashtechno/wyzely-detect"
keywords = ["object-detection", "face-detection", "wyze", "security", "yolov8", "unified-push"]
license = "MIT"
readme = "README.md"
packages = [{include = "wyzely_detect"}]
[tool.poetry.dependencies]
# Works on 3.10 and 3.11, at least in my testing
2023-10-14 02:51:16 +01:00
python = ">=3.10, <3.12"
python-dotenv = "^1.0.0"
httpx = "^0.25.0"
opencv-python = "^4.8.1.78"
ultralytics = "^8.0.190"
hjson = "^3.1.0"
numpy = "^1.23.2"
2023-10-14 02:51:16 +01:00
# https://github.com/python-poetry/poetry/issues/6409#issuecomment-1911735833
# If GPU support doesn't work, `poetry install -E cuda --with gpu` will force it to be installed from the GPU PyTorch repo
# However, PyPi's `torch` has CUDA 12.1 support by default on Linux, so in that case it should not be needed.
torch = [
{version = "^2.2.1", markers = "extra!='cuda' and platform_system=='Linux'"},
{version = "^2.2.1", source = "pytorch-cpu", markers = "extra!='cuda' and platform_system!='Linux'"},
]
# https://stackoverflow.com/a/76477590/18270659
# https://discfuss.tensorflow.org/t/tensorflow-io-gcs-filesystem-with-windows/18849/4
# https://github.com/python-poetry/poetry/issues/8271#issuecomment-1712020965
# Might be able to remove this version constraint later
2023-12-08 22:26:04 +00:00
# Working versions:
# Python version 3.10.12 and 3.10.5 both work
# CUDA version - 12.2
# cuDNN version - 8.8.1
# Installed from Nvidia website - nvidia-cuda-toolkit is not installed, but default PopOS drivers are installed
absl-py = "^2.1.0"
tensorflow = {version = "^2.13.0", markers = "extra!='cuda'"}
tensorflow-macos = { version = "^2.13.0", platform = "darwin", markers = "platform_machine=='arm64'" }
tensorflow-intel = { version = "^2.13.0", platform = "win32" }
tensorflow-io-gcs-filesystem = [
{ version = "< 0.32.0", markers = "platform_system == 'Windows'" }
]
2023-10-14 02:51:16 +01:00
deepface = "^0.0.79"
prettytable = "^3.9.0"
2024-02-11 20:27:05 +00:00
[tool.poetry.group.gpu]
optional = true
2024-02-11 20:27:05 +00:00
[tool.poetry.group.gpu.dependencies]
torch = {version = "^2.2.1", source = "pytorch-cu121", markers = "extra=='cuda'"}
2024-02-11 20:27:05 +00:00
tensorflow = {version = "^2.14.0", extras = ["and-cuda"], markers = "extra=='cuda'"}
[tool.poetry.extras]
2024-02-11 20:27:05 +00:00
# Might be better to rename this to nocpu since it's more accurate
cuda = []
[[tool.poetry.source]]
name = "pytorch-cpu"
url = "https://download.pytorch.org/whl/cpu"
priority = "explicit"
[[tool.poetry.source]]
name = "pytorch-cu121"
url = "https://download.pytorch.org/whl/cu121"
priority = "explicit"
[tool.poetry.group.dev.dependencies]
black = "^23.9.1"
ruff = "^0.0.291"
ipykernel = "^6.25.2"
2023-10-14 02:51:16 +01:00
nbconvert = "^7.9.2"
[build-system]
requires = ["poetry-core"]
build-backend = "poetry.core.masonry.api"
2023-10-14 00:16:55 +01:00
[tool.ruff]
# More than the default (88) of `black` to make comments less of a headache
2023-10-14 21:40:36 +01:00
# Where possible, `black` will attempt to format to 88 characters
# However, setting ruff to 135 will allow for longer lines that can't be auto-formatted
line-length = 135
extend-select= ["FIX002"]
2023-10-14 21:46:42 +01:00
[tool.poetry.scripts]
wyzely-detect = "wyzely_detect.__main__:main"