From 0626d21ff01f57d7e5e9b621f4db03a7d836afcf Mon Sep 17 00:00:00 2001 From: Gabriel Almeida Date: Fri, 12 May 2023 09:39:03 -0300 Subject: [PATCH] feat(pyproject.toml): add gptcache dependency to the project. --- poetry.lock | 20 +++++++++++++++++++- pyproject.toml | 1 + 2 files changed, 20 insertions(+), 1 deletion(-) diff --git a/poetry.lock b/poetry.lock index 3c772147d..2f42ef3ff 100644 --- a/poetry.lock +++ b/poetry.lock @@ -1178,6 +1178,24 @@ protobuf = ">=3.19.5,<3.20.0 || >3.20.0,<3.20.1 || >3.20.1,<4.21.1 || >4.21.1,<4 [package.extras] grpc = ["grpcio (>=1.44.0,<2.0.0dev)"] +[[package]] +name = "gptcache" +version = "0.1.23" +description = "GPTCache, a powerful caching library that can be used to speed up and lower the cost of chat applications that rely on the LLM service. GPTCache works as a memcache for AIGC applications, similar to how Redis works for traditional applications." +category = "main" +optional = false +python-versions = ">=3.8.1" +files = [ + {file = "gptcache-0.1.23-py3-none-any.whl", hash = "sha256:8bcd366e1dd5de432e113831afdea97493f090372a752a42b9ff16cb8c818635"}, + {file = "gptcache-0.1.23.tar.gz", hash = "sha256:5b5e3ef6f5df35f948bd203d1e33f3985459e60be436547529ff8b31f245238d"}, +] + +[package.dependencies] +cachetools = "*" +numpy = "*" +openai = "*" +requests = "*" + [[package]] name = "greenlet" version = "2.0.2" @@ -4822,4 +4840,4 @@ cffi = ["cffi (>=1.11)"] [metadata] lock-version = "2.0" python-versions = "^3.9" -content-hash = "79eea9f10d5eb0270db61dc96706895c52aebe9ec53cd596191ad0e107d762b5" +content-hash = "d914f734f4ff1bcbe8e678d46b20b73c3565b3af6f5dd0ac0359fae800c6bf2e" diff --git a/pyproject.toml b/pyproject.toml index d8892d4cb..1e819234a 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -50,6 +50,7 @@ pyarrow = "^11.0.0" websockets = "^11.0.2" tiktoken = "^0.3.3" wikipedia = "^1.4.0" +gptcache = "^0.1.23" [tool.poetry.group.dev.dependencies] black = "^23.1.0"