From 28fde19e8ad338f80dbf907c2ca01d2a898cabbf Mon Sep 17 00:00:00 2001 From: Gabriel Luiz Freitas Almeida Date: Sun, 2 Jul 2023 10:47:09 -0300 Subject: [PATCH] =?UTF-8?q?=F0=9F=93=A6=20chore(pyproject.toml):=20make=20?= =?UTF-8?q?huggingface-hub,=20llama-cpp-python,=20sentence-transformers,?= =?UTF-8?q?=20and=20ctransformers=20optional=20dependencies=20=F0=9F=94=80?= =?UTF-8?q?=20chore(pyproject.toml):=20update=20the=20version=20of=20huggi?= =?UTF-8?q?ngface-hub,=20llama-cpp-python,=20sentence-transformers,=20and?= =?UTF-8?q?=20ctransformers=20to=20their=20latest=20versions=20The=20huggi?= =?UTF-8?q?ngface-hub,=20llama-cpp-python,=20sentence-transformers,=20and?= =?UTF-8?q?=20ctransformers=20dependencies=20are=20now=20marked=20as=20opt?= =?UTF-8?q?ional=20in=20the=20pyproject.toml=20file.=20This=20allows=20use?= =?UTF-8?q?rs=20to=20choose=20whether=20or=20not=20to=20install=20these=20?= =?UTF-8?q?dependencies=20based=20on=20their=20specific=20needs.=20Additio?= =?UTF-8?q?nally,=20the=20versions=20of=20these=20dependencies=20have=20be?= =?UTF-8?q?en=20updated=20to=20their=20latest=20available=20versions.?= MIME-Version: 1.0 Content-Type: text/plain; charset=UTF-8 Content-Transfer-Encoding: 8bit --- poetry.lock | 34 ++++++++++++++++++---------------- pyproject.toml | 16 ++++++++++++---- 2 files changed, 30 insertions(+), 20 deletions(-) diff --git a/poetry.lock b/poetry.lock index 0760411e8..8a6aa2762 100644 --- a/poetry.lock +++ b/poetry.lock @@ -958,7 +958,7 @@ name = "ctransformers" version = "0.2.10" description = "Python bindings for the Transformer models implemented in C/C++ using GGML library." category = "main" -optional = false +optional = true python-versions = "*" files = [ {file = "ctransformers-0.2.10-py3-none-any.whl", hash = "sha256:912a80859bd252e2a389b4716d44b0663657148a85fbfbe6c5503a7ee69fd235"}, @@ -1520,7 +1520,7 @@ name = "fsspec" version = "2023.6.0" description = "File-system specification" category = "main" -optional = false +optional = true python-versions = ">=3.8" files = [ {file = "fsspec-2023.6.0-py3-none-any.whl", hash = "sha256:1cbad1faef3e391fba6dc005ae9b5bdcbf43005c9167ce78c915549c352c869a"}, @@ -2406,7 +2406,7 @@ name = "huggingface-hub" version = "0.15.1" description = "Client library to download and publish models, datasets and other repos on the huggingface.co hub" category = "main" -optional = false +optional = true python-versions = ">=3.7.0" files = [ {file = "huggingface_hub-0.15.1-py3-none-any.whl", hash = "sha256:05b0fb0abbf1f625dfee864648ac3049fe225ac4371c7bafaca0c2d3a2f83445"}, @@ -2836,7 +2836,7 @@ name = "jinja2" version = "3.1.2" description = "A very fast and expressive template engine." category = "main" -optional = false +optional = true python-versions = ">=3.7" files = [ {file = "Jinja2-3.1.2-py3-none-any.whl", hash = "sha256:6088930bfe239f0e6710546ab9c19c9ef35e29792895fed6e6e31a023a182a61"}, @@ -3039,7 +3039,7 @@ name = "llama-cpp-python" version = "0.1.55" description = "A Python wrapper for llama.cpp" category = "main" -optional = false +optional = true python-versions = ">=3.7" files = [ {file = "llama_cpp_python-0.1.55.tar.gz", hash = "sha256:1bc749f314a979c601b2dae22eb1f2d63fe791bc1237cce24d36b4f856be8ca2"}, @@ -3260,7 +3260,7 @@ name = "markupsafe" version = "2.1.3" description = "Safely add untrusted strings to HTML/XML markup." category = "main" -optional = false +optional = true python-versions = ">=3.7" files = [ {file = "MarkupSafe-2.1.3-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:cd0f502fe016460680cd20aaa5a76d241d6f35a1c3350c474bac1273803893fa"}, @@ -5933,7 +5933,7 @@ name = "safetensors" version = "0.3.1" description = "Fast and Safe Tensor serialization" category = "main" -optional = false +optional = true python-versions = "*" files = [ {file = "safetensors-0.3.1-cp310-cp310-macosx_10_11_x86_64.whl", hash = "sha256:2ae9b7dd268b4bae6624729dac86deb82104820e9786429b0583e5168db2f770"}, @@ -5994,7 +5994,7 @@ name = "scikit-learn" version = "1.3.0" description = "A set of python modules for machine learning and data mining" category = "main" -optional = false +optional = true python-versions = ">=3.8" files = [ {file = "scikit-learn-1.3.0.tar.gz", hash = "sha256:8be549886f5eda46436b6e555b0e4873b4f10aa21c07df45c4bc1735afbccd7a"}, @@ -6037,7 +6037,7 @@ name = "scipy" version = "1.11.1" description = "Fundamental algorithms for scientific computing in Python" category = "main" -optional = false +optional = true python-versions = "<3.13,>=3.9" files = [ {file = "scipy-1.11.1-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:aec8c62fbe52914f9cf28d846cf0401dd80ab80788bbab909434eb336ed07c04"}, @@ -6102,7 +6102,7 @@ name = "sentence-transformers" version = "2.2.2" description = "Multilingual text embeddings" category = "main" -optional = false +optional = true python-versions = ">=3.6.0" files = [ {file = "sentence-transformers-2.2.2.tar.gz", hash = "sha256:dbc60163b27de21076c9a30d24b5b7b6fa05141d68cf2553fa9a77bf79a29136"}, @@ -6125,7 +6125,7 @@ name = "sentencepiece" version = "0.1.99" description = "SentencePiece python wrapper" category = "main" -optional = false +optional = true python-versions = "*" files = [ {file = "sentencepiece-0.1.99-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:0eb528e70571b7c02723e5804322469b82fe7ea418c96051d0286c0fa028db73"}, @@ -6622,7 +6622,7 @@ name = "threadpoolctl" version = "3.1.0" description = "threadpoolctl" category = "main" -optional = false +optional = true python-versions = ">=3.6" files = [ {file = "threadpoolctl-3.1.0-py3-none-any.whl", hash = "sha256:8b99adda265feb6773280df41eece7b2e6561b772d21ffd52e372f999024907b"}, @@ -6771,7 +6771,7 @@ name = "torch" version = "2.0.1" description = "Tensors and Dynamic neural networks in Python with strong GPU acceleration" category = "main" -optional = false +optional = true python-versions = ">=3.8.0" files = [ {file = "torch-2.0.1-cp310-cp310-manylinux1_x86_64.whl", hash = "sha256:8ced00b3ba471856b993822508f77c98f48a458623596a4c43136158781e306a"}, @@ -6811,7 +6811,7 @@ name = "torchvision" version = "0.15.2" description = "image and video datasets and models for torch deep learning" category = "main" -optional = false +optional = true python-versions = ">=3.8" files = [ {file = "torchvision-0.15.2-cp310-cp310-macosx_10_9_x86_64.whl", hash = "sha256:7754088774e810c5672b142a45dcf20b1bd986a5a7da90f8660c43dc43fb850c"}, @@ -6908,7 +6908,7 @@ name = "transformers" version = "4.30.2" description = "State-of-the-art Machine Learning for JAX, PyTorch and TensorFlow" category = "main" -optional = false +optional = true python-versions = ">=3.7.0" files = [ {file = "transformers-4.30.2-py3-none-any.whl", hash = "sha256:c332e3a3097f9ed89ce556b403251235931c00237b8bc2d7adaa19d226c13f1d"}, @@ -7825,9 +7825,11 @@ cffi = {version = ">=1.11", markers = "platform_python_implementation == \"PyPy\ cffi = ["cffi (>=1.11)"] [extras] +all = [] deploy = ["langchain-serve"] +local = ["ctransformers", "huggingface-hub", "llama-cpp-python", "sentence-transformers"] [metadata] lock-version = "2.0" python-versions = ">=3.9,<3.11" -content-hash = "3b615d7e85c445b293174b2cb1d395e375edce224c0da51e03c3767c4892e662" +content-hash = "394bf73b8a7b70448bf8112645c3526482d733d33a829dff318d836b291a3065" diff --git a/pyproject.toml b/pyproject.toml index a9df0a27a..b0ced68a3 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -34,9 +34,9 @@ langchain = "^0.0.219" openai = "^0.27.8" pandas = "^2.0.0" chromadb = "^0.3.21" -huggingface-hub = "^0.15.0" +huggingface-hub = { version = "^0.15.0", optional = true } rich = "^13.4.2" -llama-cpp-python = "~0.1.0" +llama-cpp-python = { version = "~0.1.0", optional = true } networkx = "^3.1" unstructured = "^0.7.0" pypdf = "^3.11.0" @@ -53,8 +53,8 @@ qdrant-client = "^1.3.0" websockets = "^10.3" weaviate-client = "^3.21.0" jina = "3.15.2" -sentence-transformers = "^2.2.2" -ctransformers = "^0.2.10" +sentence-transformers = { version = "^2.2.2", optional = true } +ctransformers = { version = "^0.2.10", optional = true } cohere = "^4.11.0" python-multipart = "^0.0.6" sqlmodel = "^0.0.8" @@ -91,6 +91,14 @@ types-pyyaml = "^6.0.12.8" [tool.poetry.extras] deploy = ["langchain-serve"] +local = [ + "llama-cpp-python", + "sentence-transformers", + "ctransformers", + "huggingface-hub", +] +all = ["deploy", "local"] + [tool.pytest.ini_options] minversion = "6.0"