Support llama-index install option for upstreaming purposes (#10866)

* Support llama-index install option for upstreaming purposes

* Small fix

* Small fix
This commit is contained in:
Yuwen Hu 2024-04-23 19:08:29 +08:00 committed by GitHub
parent 21bb8bd164
commit 5c9eb5d0f5
No known key found for this signature in database
GPG key ID: B5690EEEBB952194

View file

@ -275,6 +275,15 @@ def setup_package():
] ]
all_requires += CONVERT_DEP all_requires += CONVERT_DEP
# Add internal requires for llama-index
llama_index_requires = copy.deepcopy(all_requires)
for exclude_require in ['torch', 'transformers == 4.31.0', 'tokenizers == 0.13.3']:
llama_index_requires.remove(exclude_require)
llama_index_requires += ["torch<2.2.0",
"transformers>=4.34.0,<4.39.0",
"sentence-transformers~=2.6.1"]
# Linux install with --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/ # Linux install with --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/xpu/us/
xpu_20_requires = copy.deepcopy(all_requires) xpu_20_requires = copy.deepcopy(all_requires)
xpu_20_requires.remove('torch') xpu_20_requires.remove('torch')
@ -324,7 +333,8 @@ def setup_package():
"xpu-2-0": xpu_20_requires, "xpu-2-0": xpu_20_requires,
"xpu-2-1": xpu_21_requires, "xpu-2-1": xpu_21_requires,
"serving": serving_requires, "serving": serving_requires,
"cpp": ["bigdl-core-cpp==" + CORE_XE_VERSION]}, "cpp": ["bigdl-core-cpp==" + CORE_XE_VERSION],
"llama-index": llama_index_requires}, # for internal usage when upstreaming for llama-index
classifiers=[ classifiers=[
'License :: OSI Approved :: Apache Software License', 'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3', 'Programming Language :: Python :: 3',