From d54b10e98fcf924985320c6c1d604e21d6227df0 Mon Sep 17 00:00:00 2001 From: VincyZhang Date: Sun, 12 May 2024 19:49:40 -0700 Subject: [PATCH] Update to 2.3.0 (#1539) --- .../script/unitTest/run_unit_test_neuralchat.sh | 2 +- docker/Dockerfile_chatbot | 2 +- docker/Dockerfile_code_gen | 2 +- .../huggingface/neural_speed/perplexity/requirements.txt | 2 +- examples/huggingface/neural_speed/requirements.txt | 4 ++-- .../code-generation/quantization/requirements.txt | 4 ++-- .../deployment/imagenet/vit/requirements.txt | 4 ++-- .../image-classification/quantization/requirements.txt | 4 ++-- .../fill-mask/electra_base_chinese/requirements.txt | 2 +- .../deployment/squad/bert_large/requirements.txt | 2 +- .../squad/length_adaptive_transformer/requirements.txt | 2 +- .../pytorch/question-answering/dynamic/requirements.txt | 4 ++-- .../pruning/basic_magnitude/requirements.txt | 2 +- .../pruning/longformer_triviaqa/requirements.txt | 2 +- .../pytorch/summarization/quantization/requirements.txt | 2 +- .../text-classification/cascade-models/requirements.txt | 2 +- .../emotion/distilbert_base_uncased/requirements.txt | 2 +- .../deployment/mrpc/bert_base/requirements.txt | 2 +- .../deployment/mrpc/bert_base_cased/requirements.txt | 2 +- .../deployment/mrpc/bert_mini/requirements.txt | 2 +- .../mrpc/distilbert_base_uncased/requirements.txt | 2 +- .../deployment/mrpc/roberta_base/requirements.txt | 2 +- .../deployment/sparse/bert_mini/requirements.txt | 2 +- .../sparse/distilbert_base_uncased/requirements.txt | 2 +- .../deployment/sst2/bert_mini/requirements.txt | 2 +- .../sst2/distilbert_base_uncased/requirements.txt | 2 +- .../sst2/minilm_l6_h384_uncased/requirements.txt | 2 +- .../text-classification/early-exit/requirements.txt | 2 +- .../text-classification/new_pruning/requirements.txt | 2 +- .../orchestrate_optimizations/requirements.txt | 2 +- .../pytorch/text-classification/pruning/requirements.txt | 2 +- .../quantization/ptq/requirements.txt | 2 +- .../quantization/qat/requirements.txt | 2 +- .../text-classification/quantization/requirements.txt | 2 +- .../text-embedding/deployment/mteb/bge/requirements.txt | 2 +- .../quantization/llm_quantization_recipes.md | 4 ++-- .../quantization/requirements_cpu_woq.txt | 2 +- .../text-generation/quantization/requirements_sq.txt | 4 ++-- .../deployment/stable_diffusion/requirements.txt | 2 +- .../textual-inversion/quantization/requirements.txt | 6 +++--- .../token-classification/quantization/requirements.txt | 2 +- .../pytorch/translation/quantization/requirements.txt | 2 +- .../llm/runtime/neural_speed/requirements.txt | 2 +- .../neural_chat/docker/Dockerfile | 2 +- .../neural_chat/docker/code_generation/cpu/Dockerfile | 2 +- .../neural_chat/docker/finetuning/Dockerfile | 2 +- .../neural_chat/docker/inference/README.md | 2 +- .../neural_chat/docker/text_generation/cpu/Dockerfile | 2 +- .../neural_chat/docker/tgi_serving/Dockerfile_tgi | 2 +- .../deployment/codegen/backend/xeon/tpp/README.md | 2 +- .../deployment/plugin/image2image/requirements.txt | 2 +- .../examples/finetuning/image_to_text/README.md | 2 +- .../examples/finetuning/image_to_text/requirements.txt | 2 +- .../examples/finetuning/instruction/README.md | 2 +- .../examples/finetuning/instruction/requirements.txt | 2 +- .../pipeline/plugins/retrieval/requirements.txt | 4 ++-- .../neural_chat/requirements.txt | 6 +++--- .../neural_chat/requirements_cpu.txt | 6 +++--- .../neural_chat/requirements_win.txt | 4 ++-- .../neural_chat/requirements_xpu.txt | 2 +- .../neural_chat/tests/requirements.txt | 8 ++++---- .../neural_chat/ui/gradio/basic/requirements.txt | 2 +- .../llm/evaluation/lm_eval/models/huggingface.py | 9 ++++++--- requirements-cpu.txt | 2 +- tests/requirements.txt | 6 +++--- workflows/chatbot/fine_tuning/README.md | 2 +- .../hf_finetuning_and_inference_nlp/requirements.txt | 4 ++-- 67 files changed, 93 insertions(+), 90 deletions(-) diff --git a/.github/workflows/script/unitTest/run_unit_test_neuralchat.sh b/.github/workflows/script/unitTest/run_unit_test_neuralchat.sh index 07e601a5e69..77262f47c38 100644 --- a/.github/workflows/script/unitTest/run_unit_test_neuralchat.sh +++ b/.github/workflows/script/unitTest/run_unit_test_neuralchat.sh @@ -89,7 +89,7 @@ function main() { apt-get install libsm6 libxext6 -y wget http://nz2.archive.ubuntu.com/ubuntu/pool/main/o/openssl/libssl1.1_1.1.1f-1ubuntu2.19_amd64.deb dpkg -i libssl1.1_1.1.1f-1ubuntu2.19_amd64.deb - python -m pip install --upgrade --force-reinstall torch==2.2.0 + python -m pip install --upgrade --force-reinstall torch==2.3.0 python -m pip install fastapi==0.103.2 cd ${WORKING_DIR} || exit 1 echo "test on ${test_name}" diff --git a/docker/Dockerfile_chatbot b/docker/Dockerfile_chatbot index bb3ba0a2c74..3bdb8cf9e05 100644 --- a/docker/Dockerfile_chatbot +++ b/docker/Dockerfile_chatbot @@ -56,7 +56,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi WORKDIR /intel-extension-for-transformers -RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ +RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ cd /intel-extension-for-transformers && pip install -r requirements.txt && \ python setup.py install && \ cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \ diff --git a/docker/Dockerfile_code_gen b/docker/Dockerfile_code_gen index 2f7ae579e68..f1b3ea9c8d6 100644 --- a/docker/Dockerfile_code_gen +++ b/docker/Dockerfile_code_gen @@ -49,7 +49,7 @@ COPY . /app/intel-extension-for-transformers RUN ${PYTHON} -m pip install -r requirements.txt --no-cache-dir -f https://download.pytorch.org/whl/cpu/torch_stable.html -RUN ${PYTHON} -m pip install torch==2.2.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html +RUN ${PYTHON} -m pip install torch==2.3.0+cpu -f https://download.pytorch.org/whl/cpu/torch_stable.html RUN ${PYTHON} -m pip install -r examples/huggingface/pytorch/code-generation/quantization/requirements.txt --no-cache-dir diff --git a/examples/huggingface/neural_speed/perplexity/requirements.txt b/examples/huggingface/neural_speed/perplexity/requirements.txt index ae2da4b19d9..6f747f8449e 100644 --- a/examples/huggingface/neural_speed/perplexity/requirements.txt +++ b/examples/huggingface/neural_speed/perplexity/requirements.txt @@ -1,5 +1,5 @@ --extra-index-url https://download.pytorch.org/whl/cpu -torch==2.2.0+cpu +torch==2.3.0+cpu transformers numpy sentencepiece diff --git a/examples/huggingface/neural_speed/requirements.txt b/examples/huggingface/neural_speed/requirements.txt index 250465ff365..f9750a5d066 100644 --- a/examples/huggingface/neural_speed/requirements.txt +++ b/examples/huggingface/neural_speed/requirements.txt @@ -4,8 +4,8 @@ lm-eval sentencepiece gguf --extra-index-url https://download.pytorch.org/whl/cpu -torch==2.2.0+cpu +torch==2.3.0+cpu transformers -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 tiktoken transformers_stream_generator diff --git a/examples/huggingface/pytorch/code-generation/quantization/requirements.txt b/examples/huggingface/pytorch/code-generation/quantization/requirements.txt index 1ea47d506ad..d6a40cc1746 100644 --- a/examples/huggingface/pytorch/code-generation/quantization/requirements.txt +++ b/examples/huggingface/pytorch/code-generation/quantization/requirements.txt @@ -4,12 +4,12 @@ datasets >= 2.0 protobuf sentencepiece != 0.1.92 --extra-index-url https://download.pytorch.org/whl/cpu -torch==2.2.0+cpu +torch==2.3.0+cpu peft==0.6.2 transformers >= 4.35.0 tiktoken #code_gen neural-compressor -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 optimum-intel auto-round git+https://github.com/bigcode-project/bigcode-evaluation-harness@094c7cc197d13a53c19303865e2056f1c7488ac1 diff --git a/examples/huggingface/pytorch/image-classification/deployment/imagenet/vit/requirements.txt b/examples/huggingface/pytorch/image-classification/deployment/imagenet/vit/requirements.txt index 81b825fd81e..d9e2f8e131a 100644 --- a/examples/huggingface/pytorch/image-classification/deployment/imagenet/vit/requirements.txt +++ b/examples/huggingface/pytorch/image-classification/deployment/imagenet/vit/requirements.txt @@ -4,8 +4,8 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 -torchvision==0.17.0 +torch==2.3.0 +torchvision==0.18.0 onnx>=1.12 onnxruntime==1.13.1 onnxoptimizer diff --git a/examples/huggingface/pytorch/image-classification/quantization/requirements.txt b/examples/huggingface/pytorch/image-classification/quantization/requirements.txt index 331cc5312ab..6a8c1411bec 100644 --- a/examples/huggingface/pytorch/image-classification/quantization/requirements.txt +++ b/examples/huggingface/pytorch/image-classification/quantization/requirements.txt @@ -3,8 +3,8 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 -torchvision==0.17.0 +torch==2.3.0 +torchvision==0.18.0 onnx>=1.12 onnxruntime==1.13.1 evaluate diff --git a/examples/huggingface/pytorch/language-modeling/deployment/fill-mask/electra_base_chinese/requirements.txt b/examples/huggingface/pytorch/language-modeling/deployment/fill-mask/electra_base_chinese/requirements.txt index b0b435da386..e89ebcb1eab 100644 --- a/examples/huggingface/pytorch/language-modeling/deployment/fill-mask/electra_base_chinese/requirements.txt +++ b/examples/huggingface/pytorch/language-modeling/deployment/fill-mask/electra_base_chinese/requirements.txt @@ -1,4 +1,4 @@ transformers -torch==2.2.0 +torch==2.3.0 optimum accelerate \ No newline at end of file diff --git a/examples/huggingface/pytorch/question-answering/deployment/squad/bert_large/requirements.txt b/examples/huggingface/pytorch/question-answering/deployment/squad/bert_large/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/question-answering/deployment/squad/bert_large/requirements.txt +++ b/examples/huggingface/pytorch/question-answering/deployment/squad/bert_large/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/question-answering/deployment/squad/length_adaptive_transformer/requirements.txt b/examples/huggingface/pytorch/question-answering/deployment/squad/length_adaptive_transformer/requirements.txt index 862aa6ca2ff..7e06170c1f1 100644 --- a/examples/huggingface/pytorch/question-answering/deployment/squad/length_adaptive_transformer/requirements.txt +++ b/examples/huggingface/pytorch/question-answering/deployment/squad/length_adaptive_transformer/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.12.1 optimum diff --git a/examples/huggingface/pytorch/question-answering/dynamic/requirements.txt b/examples/huggingface/pytorch/question-answering/dynamic/requirements.txt index d90b750f306..ffd6f1b1bb7 100644 --- a/examples/huggingface/pytorch/question-answering/dynamic/requirements.txt +++ b/examples/huggingface/pytorch/question-answering/dynamic/requirements.txt @@ -1,6 +1,6 @@ transformers datasets torchprofile -torch==2.2.0 -intel_extension_for_pytorch==2.2.0 +torch==2.3.0 +intel_extension_for_pytorch==2.3.0 accelerate diff --git a/examples/huggingface/pytorch/question-answering/pruning/basic_magnitude/requirements.txt b/examples/huggingface/pytorch/question-answering/pruning/basic_magnitude/requirements.txt index 90b5bf269fa..28c6c64b4b6 100644 --- a/examples/huggingface/pytorch/question-answering/pruning/basic_magnitude/requirements.txt +++ b/examples/huggingface/pytorch/question-answering/pruning/basic_magnitude/requirements.txt @@ -1,5 +1,5 @@ datasets >= 1.8.0 -torch==2.2.0 +torch==2.3.0 transformers wandb accelerate diff --git a/examples/huggingface/pytorch/question-answering/pruning/longformer_triviaqa/requirements.txt b/examples/huggingface/pytorch/question-answering/pruning/longformer_triviaqa/requirements.txt index 3ec51eef65b..84310f9ea50 100644 --- a/examples/huggingface/pytorch/question-answering/pruning/longformer_triviaqa/requirements.txt +++ b/examples/huggingface/pytorch/question-answering/pruning/longformer_triviaqa/requirements.txt @@ -1,5 +1,5 @@ accelerate datasets transformers -torch==2.2.0 +torch==2.3.0 neural-compressor==2.0 diff --git a/examples/huggingface/pytorch/summarization/quantization/requirements.txt b/examples/huggingface/pytorch/summarization/quantization/requirements.txt index 139b415893c..6441267cdcb 100644 --- a/examples/huggingface/pytorch/summarization/quantization/requirements.txt +++ b/examples/huggingface/pytorch/summarization/quantization/requirements.txt @@ -4,7 +4,7 @@ sentencepiece != 0.1.92 rouge-score nltk py7zr -torch==2.2.0 +torch==2.3.0 transformers protobuf schema diff --git a/examples/huggingface/pytorch/text-classification/cascade-models/requirements.txt b/examples/huggingface/pytorch/text-classification/cascade-models/requirements.txt index e987795b011..71ce7e588d7 100644 --- a/examples/huggingface/pytorch/text-classification/cascade-models/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/cascade-models/requirements.txt @@ -1,4 +1,4 @@ -torch==2.2.0 +torch==2.3.0 numpy transformers datasets diff --git a/examples/huggingface/pytorch/text-classification/deployment/emotion/distilbert_base_uncased/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/emotion/distilbert_base_uncased/requirements.txt index eda685151bb..6cf73c3deae 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/emotion/distilbert_base_uncased/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/emotion/distilbert_base_uncased/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base_cased/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base_cased/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base_cased/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_base_cased/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_mini/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_mini/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_mini/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/mrpc/bert_mini/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/mrpc/distilbert_base_uncased/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/mrpc/distilbert_base_uncased/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/mrpc/distilbert_base_uncased/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/mrpc/distilbert_base_uncased/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/mrpc/roberta_base/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/mrpc/roberta_base/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/mrpc/roberta_base/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/mrpc/roberta_base/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/sparse/bert_mini/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/sparse/bert_mini/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/sparse/bert_mini/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/sparse/bert_mini/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/sparse/distilbert_base_uncased/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/sparse/distilbert_base_uncased/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/sparse/distilbert_base_uncased/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/sparse/distilbert_base_uncased/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/sst2/bert_mini/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/sst2/bert_mini/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/sst2/bert_mini/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/sst2/bert_mini/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/sst2/distilbert_base_uncased/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/sst2/distilbert_base_uncased/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/sst2/distilbert_base_uncased/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/sst2/distilbert_base_uncased/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/deployment/sst2/minilm_l6_h384_uncased/requirements.txt b/examples/huggingface/pytorch/text-classification/deployment/sst2/minilm_l6_h384_uncased/requirements.txt index 2bfd563ca14..94b03297a4a 100644 --- a/examples/huggingface/pytorch/text-classification/deployment/sst2/minilm_l6_h384_uncased/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/deployment/sst2/minilm_l6_h384_uncased/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 optimum diff --git a/examples/huggingface/pytorch/text-classification/early-exit/requirements.txt b/examples/huggingface/pytorch/text-classification/early-exit/requirements.txt index 4e6b54440dc..31fc7769b30 100644 --- a/examples/huggingface/pytorch/text-classification/early-exit/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/early-exit/requirements.txt @@ -1,4 +1,4 @@ -torch==2.2.0 +torch==2.3.0 transformers datasets allennlp diff --git a/examples/huggingface/pytorch/text-classification/new_pruning/requirements.txt b/examples/huggingface/pytorch/text-classification/new_pruning/requirements.txt index 8160dbc4578..7e0efd19de8 100644 --- a/examples/huggingface/pytorch/text-classification/new_pruning/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/new_pruning/requirements.txt @@ -6,5 +6,5 @@ sentencepiece scipy scikit-learn protobuf -torch==2.2.0 +torch==2.3.0 evaluate diff --git a/examples/huggingface/pytorch/text-classification/orchestrate_optimizations/requirements.txt b/examples/huggingface/pytorch/text-classification/orchestrate_optimizations/requirements.txt index 49a8250f2f3..65fe55f0315 100644 --- a/examples/huggingface/pytorch/text-classification/orchestrate_optimizations/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/orchestrate_optimizations/requirements.txt @@ -1,5 +1,5 @@ accelerate -torch==2.2.0 +torch==2.3.0 datasets >= 1.1.3 sentencepiece != 0.1.92 transformers diff --git a/examples/huggingface/pytorch/text-classification/pruning/requirements.txt b/examples/huggingface/pytorch/text-classification/pruning/requirements.txt index f94b95c0902..54bf0391a83 100644 --- a/examples/huggingface/pytorch/text-classification/pruning/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/pruning/requirements.txt @@ -2,6 +2,6 @@ accelerate datasets >= 1.1.3 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 transformers wandb diff --git a/examples/huggingface/pytorch/text-classification/quantization/ptq/requirements.txt b/examples/huggingface/pytorch/text-classification/quantization/ptq/requirements.txt index f94b95c0902..54bf0391a83 100644 --- a/examples/huggingface/pytorch/text-classification/quantization/ptq/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/quantization/ptq/requirements.txt @@ -2,6 +2,6 @@ accelerate datasets >= 1.1.3 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 transformers wandb diff --git a/examples/huggingface/pytorch/text-classification/quantization/qat/requirements.txt b/examples/huggingface/pytorch/text-classification/quantization/qat/requirements.txt index f94b95c0902..54bf0391a83 100644 --- a/examples/huggingface/pytorch/text-classification/quantization/qat/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/quantization/qat/requirements.txt @@ -2,6 +2,6 @@ accelerate datasets >= 1.1.3 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 transformers wandb diff --git a/examples/huggingface/pytorch/text-classification/quantization/requirements.txt b/examples/huggingface/pytorch/text-classification/quantization/requirements.txt index f94b95c0902..54bf0391a83 100644 --- a/examples/huggingface/pytorch/text-classification/quantization/requirements.txt +++ b/examples/huggingface/pytorch/text-classification/quantization/requirements.txt @@ -2,6 +2,6 @@ accelerate datasets >= 1.1.3 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 transformers wandb diff --git a/examples/huggingface/pytorch/text-embedding/deployment/mteb/bge/requirements.txt b/examples/huggingface/pytorch/text-embedding/deployment/mteb/bge/requirements.txt index 44b6305238c..9c501e67d78 100644 --- a/examples/huggingface/pytorch/text-embedding/deployment/mteb/bge/requirements.txt +++ b/examples/huggingface/pytorch/text-embedding/deployment/mteb/bge/requirements.txt @@ -3,7 +3,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx onnxruntime==1.13.1 mteb==1.1.1 diff --git a/examples/huggingface/pytorch/text-generation/quantization/llm_quantization_recipes.md b/examples/huggingface/pytorch/text-generation/quantization/llm_quantization_recipes.md index 504c4df0159..0af551db7a2 100644 --- a/examples/huggingface/pytorch/text-generation/quantization/llm_quantization_recipes.md +++ b/examples/huggingface/pytorch/text-generation/quantization/llm_quantization_recipes.md @@ -42,8 +42,8 @@ cd examples/huggingface/pytorch/text-generation/quantization pip install -r requirements.txt pip install neural-compressor==2.5 pip install transformers==4.35.2 -pip install torch==2.2.0+cpu --index-url https://download.pytorch.org/whl/cpu -pip install intel-extension-for-pytorch==2.2.0 +pip install torch==2.3.0+cpu --index-url https://download.pytorch.org/whl/cpu +pip install intel-extension-for-pytorch==2.3.0 ``` # Run Quantization and evaluate INT8 accuracy diff --git a/examples/huggingface/pytorch/text-generation/quantization/requirements_cpu_woq.txt b/examples/huggingface/pytorch/text-generation/quantization/requirements_cpu_woq.txt index 69b77328457..acac73647c4 100644 --- a/examples/huggingface/pytorch/text-generation/quantization/requirements_cpu_woq.txt +++ b/examples/huggingface/pytorch/text-generation/quantization/requirements_cpu_woq.txt @@ -4,7 +4,7 @@ peft protobuf sentencepiece != 0.1.92 --extra-index-url https://download.pytorch.org/whl/cpu -torch==2.2.0+cpu +torch==2.3.0+cpu transformers bitsandbytes #baichuan transformers_stream_generator diff --git a/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt b/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt index f5b9b6cb77c..07817413f55 100644 --- a/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt +++ b/examples/huggingface/pytorch/text-generation/quantization/requirements_sq.txt @@ -4,9 +4,9 @@ peft protobuf sentencepiece != 0.1.92 --extra-index-url https://download.pytorch.org/whl/cpu -torch==2.2.0+cpu +torch==2.3.0+cpu transformers -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 git+https://github.com/huggingface/optimum.git@e38d40ad220a180213f99b1d93d0407a826c326d optimum-intel bitsandbytes #baichuan diff --git a/examples/huggingface/pytorch/text-to-image/deployment/stable_diffusion/requirements.txt b/examples/huggingface/pytorch/text-to-image/deployment/stable_diffusion/requirements.txt index cf78dc7c6f2..7295f511f4f 100644 --- a/examples/huggingface/pytorch/text-to-image/deployment/stable_diffusion/requirements.txt +++ b/examples/huggingface/pytorch/text-to-image/deployment/stable_diffusion/requirements.txt @@ -4,7 +4,7 @@ accelerate datasets >= 1.8.0 sentencepiece != 0.1.92 protobuf -torch==2.2.0 +torch==2.3.0 onnx>=1.12 onnxruntime==1.13.1 diffusers==0.12.1 diff --git a/examples/huggingface/pytorch/textual-inversion/quantization/requirements.txt b/examples/huggingface/pytorch/textual-inversion/quantization/requirements.txt index f7ca444b7ed..2610ad450dd 100644 --- a/examples/huggingface/pytorch/textual-inversion/quantization/requirements.txt +++ b/examples/huggingface/pytorch/textual-inversion/quantization/requirements.txt @@ -1,9 +1,9 @@ diffusers==0.4.1 accelerate -torch==2.2.0 -torchvision==0.17.0 +torch==2.3.0 +torchvision==0.18.0 transformers ftfy tensorboard modelcards -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 diff --git a/examples/huggingface/pytorch/token-classification/quantization/requirements.txt b/examples/huggingface/pytorch/token-classification/quantization/requirements.txt index 8c34c7c9a2e..43c85be5bb4 100644 --- a/examples/huggingface/pytorch/token-classification/quantization/requirements.txt +++ b/examples/huggingface/pytorch/token-classification/quantization/requirements.txt @@ -1,6 +1,6 @@ accelerate seqeval datasets >= 1.1.3 -torch==2.2.0 +torch==2.3.0 transformers wandb diff --git a/examples/huggingface/pytorch/translation/quantization/requirements.txt b/examples/huggingface/pytorch/translation/quantization/requirements.txt index e0ddb7139a9..ab79ba6aef6 100644 --- a/examples/huggingface/pytorch/translation/quantization/requirements.txt +++ b/examples/huggingface/pytorch/translation/quantization/requirements.txt @@ -4,5 +4,5 @@ sentencepiece != 0.1.92 protobuf sacrebleu >= 1.4.12 py7zr -torch==2.2.0 +torch==2.3.0 transformers diff --git a/intel_extension_for_transformers/llm/runtime/neural_speed/requirements.txt b/intel_extension_for_transformers/llm/runtime/neural_speed/requirements.txt index 3a4669658df..ffc41b33119 100644 --- a/intel_extension_for_transformers/llm/runtime/neural_speed/requirements.txt +++ b/intel_extension_for_transformers/llm/runtime/neural_speed/requirements.txt @@ -12,6 +12,6 @@ protobuf<3.20 py-cpuinfo sentencepiece tiktoken -torch==2.2.0+cpu +torch==2.3.0+cpu transformers transformers_stream_generator diff --git a/intel_extension_for_transformers/neural_chat/docker/Dockerfile b/intel_extension_for_transformers/neural_chat/docker/Dockerfile index 7e4d25e59fd..7cc1e6687b6 100644 --- a/intel_extension_for_transformers/neural_chat/docker/Dockerfile +++ b/intel_extension_for_transformers/neural_chat/docker/Dockerfile @@ -56,7 +56,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi WORKDIR /intel-extension-for-transformers -RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ +RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ cd /intel-extension-for-transformers && pip install schema==0.7.5 numpy==1.26.4 && \ python setup.py install && \ cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \ diff --git a/intel_extension_for_transformers/neural_chat/docker/code_generation/cpu/Dockerfile b/intel_extension_for_transformers/neural_chat/docker/code_generation/cpu/Dockerfile index ef17692745c..a1d2d9452e3 100644 --- a/intel_extension_for_transformers/neural_chat/docker/code_generation/cpu/Dockerfile +++ b/intel_extension_for_transformers/neural_chat/docker/code_generation/cpu/Dockerfile @@ -45,7 +45,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi WORKDIR /intel-extension-for-transformers -RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ +RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ cd /intel-extension-for-transformers && pip install -r requirements.txt && python setup.py install && \ cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \ cd /intel-extension-for-transformers/intel_extension_for_transformers/neural_chat && pip install -r requirements_cpu.txt && \ diff --git a/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile b/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile index d8edcea4780..14f700fd605 100644 --- a/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile +++ b/intel_extension_for_transformers/neural_chat/docker/finetuning/Dockerfile @@ -50,7 +50,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi WORKDIR /intel-extension-for-transformers -RUN pip install --no-cache-dir oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ +RUN pip install --no-cache-dir oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ cd /intel-extension-for-transformers && pip install --no-cache-dir -r requirements.txt && \ python setup.py install && \ cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install --no-cache-dir -r requirements.txt && \ diff --git a/intel_extension_for_transformers/neural_chat/docker/inference/README.md b/intel_extension_for_transformers/neural_chat/docker/inference/README.md index 4bd60a71df6..40650c68341 100644 --- a/intel_extension_for_transformers/neural_chat/docker/inference/README.md +++ b/intel_extension_for_transformers/neural_chat/docker/inference/README.md @@ -82,7 +82,7 @@ docker run -it --runtime=habana --name="chatbot" -e HABANA_VISIBLE_DEVICES=all - docker exec -it chatbot /bin/bash ## run inference unittest pip install -r pipeline/plugins/audio/requirements.txt -pip install --upgrade --force-reinstall torch==2.2.0 +pip install --upgrade --force-reinstall torch==2.3.0 cd tests/ci/api python test_inference.py diff --git a/intel_extension_for_transformers/neural_chat/docker/text_generation/cpu/Dockerfile b/intel_extension_for_transformers/neural_chat/docker/text_generation/cpu/Dockerfile index e63ea846f6a..da389b54b05 100644 --- a/intel_extension_for_transformers/neural_chat/docker/text_generation/cpu/Dockerfile +++ b/intel_extension_for_transformers/neural_chat/docker/text_generation/cpu/Dockerfile @@ -45,7 +45,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi WORKDIR /intel-extension-for-transformers -RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ +RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ cd /intel-extension-for-transformers && pip install -r requirements.txt && python setup.py install && \ cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \ cd /intel-extension-for-transformers/intel_extension_for_transformers/neural_chat && pip install -r requirements_cpu.txt && \ diff --git a/intel_extension_for_transformers/neural_chat/docker/tgi_serving/Dockerfile_tgi b/intel_extension_for_transformers/neural_chat/docker/tgi_serving/Dockerfile_tgi index f67278779d7..a51ca5995c8 100644 --- a/intel_extension_for_transformers/neural_chat/docker/tgi_serving/Dockerfile_tgi +++ b/intel_extension_for_transformers/neural_chat/docker/tgi_serving/Dockerfile_tgi @@ -70,7 +70,7 @@ COPY ${REPO_PATH} /intel-extension-for-transformers RUN if [ "$REPO_PATH" == "" ]; then rm -rf intel-extension-for-transformers/* && rm -rf intel-extension-for-transformers/.* ; git clone --single-branch --branch=${ITREX_VER} ${REPO} intel-extension-for-transformers ; fi WORKDIR /intel-extension-for-transformers -RUN pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ +RUN pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ && \ cd /intel-extension-for-transformers && pip install -r requirements.txt && \ pip install -v . && \ cd ./intel_extension_for_transformers/neural_chat/examples/finetuning/instruction && pip install -r requirements.txt && \ diff --git a/intel_extension_for_transformers/neural_chat/examples/deployment/codegen/backend/xeon/tpp/README.md b/intel_extension_for_transformers/neural_chat/examples/deployment/codegen/backend/xeon/tpp/README.md index c7e6d6ea092..7833bf222cf 100644 --- a/intel_extension_for_transformers/neural_chat/examples/deployment/codegen/backend/xeon/tpp/README.md +++ b/intel_extension_for_transformers/neural_chat/examples/deployment/codegen/backend/xeon/tpp/README.md @@ -42,7 +42,7 @@ pip install -r ../../../../../../pipeline/plugins/retrieval/requirements.txt pip uninstall torch torchvision torchaudio intel-extension-for-pytorch -y python -m pip install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cpu python -m pip install intel-extension-for-pytorch -python -m pip install oneccl_bind_pt --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ +python -m pip install oneccl_bind_pt==2.3.0 --extra-index-url https://pytorch-extension.intel.com/release-whl/stable/cpu/us/ pip install transformers==4.31.0 # need to downgrade transformers to 4.31.0 for LLAMA ``` diff --git a/intel_extension_for_transformers/neural_chat/examples/deployment/plugin/image2image/requirements.txt b/intel_extension_for_transformers/neural_chat/examples/deployment/plugin/image2image/requirements.txt index 64adc5235f1..295a3782b9a 100644 --- a/intel_extension_for_transformers/neural_chat/examples/deployment/plugin/image2image/requirements.txt +++ b/intel_extension_for_transformers/neural_chat/examples/deployment/plugin/image2image/requirements.txt @@ -8,5 +8,5 @@ optimum protobuf pytorch_fid sentencepiece != 0.1.92 -torch==2.2.0 +torch==2.3.0 transformers diff --git a/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/README.md b/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/README.md index 8af5e24f46d..324c8e9087a 100644 --- a/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/README.md +++ b/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/README.md @@ -12,7 +12,7 @@ Recommend python 3.9 or higher version. pip install -r requirements.txt pip install transformers==4.34.1 # To use ccl as the distributed backend in distributed training on CPU requires to install below requirement. -python -m pip install oneccl_bind_pt==2.2.0 -f https://developer.intel.com/ipex-whl-stable-cpu +python -m pip install oneccl_bind_pt==2.3.0 -f https://developer.intel.com/ipex-whl-stable-cpu ``` >**Note**: Suggest using transformers no higher than 4.34.1 diff --git a/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/requirements.txt b/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/requirements.txt index 5bd0dd28b8a..8ed10d8eeaa 100644 --- a/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/requirements.txt +++ b/intel_extension_for_transformers/neural_chat/examples/finetuning/image_to_text/requirements.txt @@ -9,7 +9,7 @@ python-multipart rouge_score sentencepiece shortuuid -torch==2.2.0 +torch==2.3.0 transformers uvicorn yacs diff --git a/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/README.md b/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/README.md index e32fc6a40b6..9cbe651d29a 100644 --- a/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/README.md +++ b/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/README.md @@ -23,7 +23,7 @@ Recommend python 3.9 or higher version. pip install -r requirements.txt pip install transformers==4.34.1 # To use ccl as the distributed backend in distributed training on CPU requires to install below requirement. -python -m pip install oneccl_bind_pt==2.2.0 -f https://developer.intel.com/ipex-whl-stable-cpu +python -m pip install oneccl_bind_pt==2.3.0 -f https://developer.intel.com/ipex-whl-stable-cpu ``` >**Note**: Suggest using transformers no higher than 4.34.1 diff --git a/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/requirements.txt b/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/requirements.txt index 5bd0dd28b8a..8ed10d8eeaa 100644 --- a/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/requirements.txt +++ b/intel_extension_for_transformers/neural_chat/examples/finetuning/instruction/requirements.txt @@ -9,7 +9,7 @@ python-multipart rouge_score sentencepiece shortuuid -torch==2.2.0 +torch==2.3.0 transformers uvicorn yacs diff --git a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/requirements.txt b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/requirements.txt index c223c306ea0..538d3c32abd 100644 --- a/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/requirements.txt +++ b/intel_extension_for_transformers/neural_chat/pipeline/plugins/retrieval/requirements.txt @@ -15,6 +15,6 @@ qdrant-client==1.8.2 rank_bm25 scikit-learn sentence-transformers==2.3.1 -torch==2.2.0 -torchaudio==2.2.0 +torch==2.3.0 +torchaudio==2.3.0 unstructured[all-docs] diff --git a/intel_extension_for_transformers/neural_chat/requirements.txt b/intel_extension_for_transformers/neural_chat/requirements.txt index 9e7b18a1c7f..2953d834f2e 100644 --- a/intel_extension_for_transformers/neural_chat/requirements.txt +++ b/intel_extension_for_transformers/neural_chat/requirements.txt @@ -5,7 +5,7 @@ evaluate fastapi fschat==0.2.35 huggingface_hub -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 lm-eval neural-compressor neural_speed==1.0a0 @@ -22,8 +22,8 @@ sacremoses shortuuid starlette tensorflow>=2.13.0 -torch==2.2.0 -torchaudio==2.2.0 +torch==2.3.0 +torchaudio==2.3.0 transformers>=4.35.2 transformers_stream_generator uvicorn diff --git a/intel_extension_for_transformers/neural_chat/requirements_cpu.txt b/intel_extension_for_transformers/neural_chat/requirements_cpu.txt index 709426d1165..267b596c280 100644 --- a/intel_extension_for_transformers/neural_chat/requirements_cpu.txt +++ b/intel_extension_for_transformers/neural_chat/requirements_cpu.txt @@ -5,7 +5,7 @@ evaluate fastapi fschat==0.2.32 huggingface_hub -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 lm-eval neural-compressor neural_speed==1.0a0 @@ -21,8 +21,8 @@ sacremoses shortuuid starlette tiktoken==0.4.0 -torch==2.2.0 -torchaudio==2.2.0 +torch==2.3.0 +torchaudio==2.3.0 transformers>=4.35.2 transformers_stream_generator uvicorn diff --git a/intel_extension_for_transformers/neural_chat/requirements_win.txt b/intel_extension_for_transformers/neural_chat/requirements_win.txt index f8137fde45f..c417c5ca01a 100644 --- a/intel_extension_for_transformers/neural_chat/requirements_win.txt +++ b/intel_extension_for_transformers/neural_chat/requirements_win.txt @@ -19,8 +19,8 @@ rouge_score sacremoses shortuuid starlette -torch==2.2.0 -torchaudio==2.2.0 +torch==2.3.0 +torchaudio==2.3.0 transformers transformers_stream_generator uvicorn diff --git a/intel_extension_for_transformers/neural_chat/requirements_xpu.txt b/intel_extension_for_transformers/neural_chat/requirements_xpu.txt index 507b9872332..9e07bc8e8f0 100644 --- a/intel_extension_for_transformers/neural_chat/requirements_xpu.txt +++ b/intel_extension_for_transformers/neural_chat/requirements_xpu.txt @@ -5,7 +5,7 @@ evaluate fastapi fschat==0.2.35 huggingface_hub -intel-extension-for-pytorch==2.1.10+xpu +intel-extension-for-pytorch==2.1.30+xpu neural-compressor numpy==1.23.5 pydantic==1.10.13 diff --git a/intel_extension_for_transformers/neural_chat/tests/requirements.txt b/intel_extension_for_transformers/neural_chat/tests/requirements.txt index a7897ee1ba9..a1d17030e6d 100644 --- a/intel_extension_for_transformers/neural_chat/tests/requirements.txt +++ b/intel_extension_for_transformers/neural_chat/tests/requirements.txt @@ -27,7 +27,7 @@ imageio==2.19.3 imageio-ffmpeg==0.4.7 inflect InstructorEmbedding -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 jaconv jieba joblib==1.2.0 @@ -79,9 +79,9 @@ spacy==3.7.4 speechbrain==0.5.15 starlette tiktoken==0.4.0 -torch==2.2.0 -torchaudio==2.2.0 -torchvision==0.17.0 +torch==2.3.0 +torchaudio==2.3.0 +torchvision==0.18.0 tqdm transformers==4.38.0 transformers_stream_generator diff --git a/intel_extension_for_transformers/neural_chat/ui/gradio/basic/requirements.txt b/intel_extension_for_transformers/neural_chat/ui/gradio/basic/requirements.txt index ab18c3a46a7..9a6771e78d9 100644 --- a/intel_extension_for_transformers/neural_chat/ui/gradio/basic/requirements.txt +++ b/intel_extension_for_transformers/neural_chat/ui/gradio/basic/requirements.txt @@ -8,5 +8,5 @@ openai pip pydantic==1.10.13 requests -torch==2.2.0 +torch==2.3.0 transformers diff --git a/intel_extension_for_transformers/transformers/llm/evaluation/lm_eval/models/huggingface.py b/intel_extension_for_transformers/transformers/llm/evaluation/lm_eval/models/huggingface.py index 6abdba8ea5a..42541846c58 100644 --- a/intel_extension_for_transformers/transformers/llm/evaluation/lm_eval/models/huggingface.py +++ b/intel_extension_for_transformers/transformers/llm/evaluation/lm_eval/models/huggingface.py @@ -648,7 +648,8 @@ def _create_model( key_value_input_names = [key for key in inputs_names if (".key" in key) or (".value" in key)] use_cache = len(key_value_input_names) > 0 - self._model = ORTModelForCausalLM(session[0], # pylint: disable=E1121 + # pylint: disable=E1121,E1124 + self._model = ORTModelForCausalLM(session[0], model_config, pretrained, use_cache=True if use_cache else False, @@ -673,10 +674,12 @@ def _create_model( sessions[1], use_cache=True) else: - sessions = ORTModelForCausalLM.load_model( # pylint: disable=E1123 + # pylint: disable=E1123,E1124 + sessions = ORTModelForCausalLM.load_model( os.path.join(pretrained, "decoder_model.onnx"), session_options=sess_options) - self._model = ORTModelForCausalLM(sessions[0], # pylint: disable=E1121 + # pylint: disable=E1121,E1124 + self._model = ORTModelForCausalLM(sessions[0], model_config, pretrained, use_cache=False, diff --git a/requirements-cpu.txt b/requirements-cpu.txt index 434b5fe2610..061bce7355e 100644 --- a/requirements-cpu.txt +++ b/requirements-cpu.txt @@ -1,4 +1,4 @@ --extra-index-url https://download.pytorch.org/whl/cpu cmake ninja -torch==2.2.0+cpu +torch==2.3.0+cpu diff --git a/tests/requirements.txt b/tests/requirements.txt index eff01799127..fa7ebb7dcf7 100644 --- a/tests/requirements.txt +++ b/tests/requirements.txt @@ -9,7 +9,7 @@ evaluate gguf git+https://github.com/intel/neural-compressor.git git+https://github.com/intel/neural-speed.git -intel-extension-for-pytorch==2.2.0 +intel-extension-for-pytorch==2.3.0 intel-tensorflow==2.14.0 lm-eval==0.4.2 mlflow @@ -23,9 +23,9 @@ sacremoses sentencepiece != 0.1.92 tiktoken tokenizers==0.15.2 -torch==2.2.0+cpu +torch==2.3.0+cpu torchprofile -torchvision==0.17.0+cpu +torchvision==0.18.0+cpu transformers==4.37.2 transformers_stream_generator tyro diff --git a/workflows/chatbot/fine_tuning/README.md b/workflows/chatbot/fine_tuning/README.md index e6f62387eac..6d9fabb072c 100644 --- a/workflows/chatbot/fine_tuning/README.md +++ b/workflows/chatbot/fine_tuning/README.md @@ -18,7 +18,7 @@ Recommend python 3.9 or higher version. ```shell pip install -r requirements.txt # To use ccl as the distributed backend in distributed training on CPU requires to install below requirement. -python -m pip install oneccl_bind_pt==2.2.0 -f https://developer.intel.com/ipex-whl-stable-cpu +python -m pip install oneccl_bind_pt==2.3.0 -f https://developer.intel.com/ipex-whl-stable-cpu ``` ## 2. Prepare the Model diff --git a/workflows/hf_finetuning_and_inference_nlp/requirements.txt b/workflows/hf_finetuning_and_inference_nlp/requirements.txt index 8725b721408..a267e1bba18 100644 --- a/workflows/hf_finetuning_and_inference_nlp/requirements.txt +++ b/workflows/hf_finetuning_and_inference_nlp/requirements.txt @@ -2,7 +2,7 @@ accelerate==0.21.0 datasets==2.11.0 intel-extension-for-transformers==1.0.0 -intel_extension_for_pytorch==2.2.0 +intel_extension_for_pytorch==2.3.0 neural-compressor==2.1 -torch==2.2.0 +torch==2.3.0 transformers