-
Notifications
You must be signed in to change notification settings - Fork 3
/
Copy pathpyproject.toml
108 lines (101 loc) · 3.39 KB
/
pyproject.toml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
[build-system]
requires = ["pdm-backend"]
build-backend = "pdm.backend"
[project]
name = "unstract-adapters"
dynamic = ["version"]
description = "Unstract interface for LLMs, Embeddings and VectorDBs"
dependencies = [
"llama-index==0.10.38",
"llama-index-embeddings-google==0.1.5",
"llama-index-embeddings-azure-openai==0.1.6",
# llama-index itself has a dependency on open-ai.
# Hence, not explicitly pinning any version
# "llama-index-embeddings-openai==0.1.7",
# Disabling Hugging Face & FastEmbed to
# keep the image size under check
# "llama-index-embeddings-huggingface==0.2.0",
# Disabling fast embed due to high processing power
# "llama-index-embeddings-fastembed==0.1.4",
"llama-index-embeddings-azure-openai==0.1.6",
"llama-index-embeddings-ollama==0.1.2",
"llama-index-vector-stores-postgres==0.1.3",
# Including Supabase conflicts with postgres on pg-vector.
# Hence, commenting it out at the moment
# "llama-index-vector-stores-supabase==0.1.3",
"llama-index-vector-stores-milvus==0.1.18",
"llama-index-vector-stores-weaviate==0.1.4",
"llama-index-vector-stores-pinecone==0.1.4",
"llama-index-vector-stores-qdrant==0.2.8",
"llama-index-llms-palm==0.1.5",
"llama-index-llms-mistralai==0.1.10",
"llama-index-llms-anyscale==0.1.3",
"llama-index-llms-anthropic==0.1.11",
# Llama-index 0.10.x implcitly includes openai for LLM
"llama-index-llms-azure-openai==0.1.5",
"llama-index-llms-vertex==0.1.8",
"llama-index-llms-replicate==0.1.3",
"llama-index-llms-ollama==0.1.3",
# OCR
"filetype~=1.2.0",
# Others
# For singleton classes
"singleton-decorator~=1.0.0",
# For Llama Parse X2Text
"llama-parse==0.4.1",
"httpx>=0.25.2",
]
readme = "README.md"
urls = { Homepage = "https://unstract.com", "Release notes" = "https://github.com/Zipstack/unstract-adapters/releases", Source = "https://github.com/Zipstack/unstract-adapters" }
license = {text = "AGPL v3"}
authors = [
{name = "Zipstack Inc.", email = "[email protected]"},
]
requires-python = ">=3.9,<3.12"
classifiers = [
"Development Status :: 4 - Beta",
"Intended Audience :: Developers",
"License :: OSI Approved :: GNU Affero General Public License v3",
"Operating System :: POSIX :: Linux",
"Operating System :: MacOS :: MacOS X",
"Programming Language :: Python",
"Programming Language :: Python :: 3.9",
"Topic :: Scientific/Engineering",
"Topic :: Software Development :: Libraries :: Python Modules",
]
[tool.pdm.version]
source = "file"
path = "src/unstract/adapters/__init__.py"
[tool.pdm.dev-dependencies]
lint = [
"absolufy-imports~=0.3.1",
"autopep8~=2.0.2",
"black~=24.3.0",
"docutils~=0.20.1",
"flake8~=7.0.0",
"flake8-pyproject~=1.2.2",
"isort~=5.13.2",
"mypy~=1.9.0",
"pre-commit~=3.6.2",
"types-PyMySQL~=1.1.0.1",
"types-pyOpenSSL~=24.0.0.20240311",
"types-PyYAML~=6.0.12.12",
"types-redis~=4.6.0.3",
"types-requests~=2.31.0.6",
"types-tzlocal~=5.1.0.1",
"yamllint>=1.35.1",
]
[tool.isort]
line_length = 88
multi_line_output = 3
include_trailing_comma = true
force_grid_wrap = 0
use_parentheses = true
ensure_newline_before_comments = true
profile = "black"
[tool.pdm.build]
includes = ["src"]
package-dir = "src"
# source-includes = ["tests"]
[tool.pdm.resolution.overrides]
grpcio = ">=1.62.1"