Skip to content

Commit 7e582be

Browse files
authored
Merge branch 'master' into move_export
2 parents 0619f7f + 455f1e1 commit 7e582be

File tree

36 files changed

+595
-917
lines changed

36 files changed

+595
-917
lines changed

.azure-pipelines/model-test-3x.yml

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -14,6 +14,7 @@ pr:
1414
- setup.py
1515
- requirements_pt.txt
1616
- .azure-pipelines/scripts/models
17+
- .azure-pipelines/model-test-3x.yml
1718

1819
variables:
1920
OUT_SCRIPT_PATH: $(Build.SourcesDirectory)/.azure-pipelines/scripts/models
@@ -35,7 +36,7 @@ parameters:
3536
stages:
3637
- stage: PyTorchModels
3738
displayName: Run PyTorch Model
38-
pool: MODEL_PERF_TEST
39+
pool: ICX-16C
3940
dependsOn: []
4041
condition: and(succeeded(), eq('${{ parameters.PyTorch_Model_3X }}', 'true'))
4142
jobs:
@@ -60,7 +61,7 @@ stages:
6061
- script: |
6162
echo ${BUILD_SOURCESDIRECTORY}
6263
rm -fr ${BUILD_SOURCESDIRECTORY} || sudo rm -fr ${BUILD_SOURCESDIRECTORY} || true
63-
echo y | docker system prune
64+
echo y | docker system prune --all
6465
displayName: "Clean workspace"
6566
- checkout: self
6667
clean: true

.azure-pipelines/model-test.yml

Lines changed: 7 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -13,6 +13,7 @@ pr:
1313
- requirements.txt
1414
- .azure-pipelines/scripts/models
1515
- examples/tensorflow/oob_models/quantization/ptq
16+
- .azure-pipelines/model-test.yml
1617
exclude:
1718
- test
1819
- neural_compressor/common
@@ -49,10 +50,10 @@ parameters:
4950
default:
5051
- resnet50v1.5
5152
- ssd_resnet50_v1
52-
- ssd_mobilenet_v1_ckpt
53-
- inception_v1
54-
- darknet19
55-
- resnet-101
53+
# - ssd_mobilenet_v1_ckpt
54+
# - inception_v1
55+
# - darknet19
56+
# - resnet-101
5657
- name: PyTorchModelList
5758
type: object
5859
default:
@@ -70,7 +71,7 @@ parameters:
7071
stages:
7172
- stage: TensorFlowModels
7273
displayName: Run TensorFlow Model
73-
pool: MODEL_PERF_TEST_TF
74+
pool: MODEL_PERF_TEST
7475
dependsOn: []
7576
condition: and(succeeded(), eq('${{ parameters.TensorFlow_Model }}', 'true'))
7677
jobs:
@@ -139,7 +140,7 @@ stages:
139140
- script: |
140141
echo ${BUILD_SOURCESDIRECTORY}
141142
rm -fr ${BUILD_SOURCESDIRECTORY} || sudo rm -fr ${BUILD_SOURCESDIRECTORY} || true
142-
echo y | docker system prune
143+
echo y | docker system prune --all
143144
displayName: "Clean workspace"
144145
- checkout: self
145146
clean: true

.azure-pipelines/scripts/fwk_version.sh

Lines changed: 8 additions & 8 deletions
Original file line numberDiff line numberDiff line change
@@ -4,21 +4,21 @@ echo "export FWs version..."
44
test_mode=$1
55

66
if [ "$test_mode" == "coverage" ] || [ "$test_mode" == "latest" ]; then
7+
export tensorflow_version='2.15.0-official'
8+
export pytorch_version='2.3.0+cpu'
9+
export torchvision_version='0.18.0+cpu'
10+
export ipex_version='2.3.0+cpu'
11+
export onnx_version='1.16.0'
12+
export onnxruntime_version='1.18.0'
13+
export mxnet_version='1.9.1'
14+
else
715
export tensorflow_version='2.15.0-official'
816
export pytorch_version='2.2.1+cpu'
917
export torchvision_version='0.17.1+cpu'
1018
export ipex_version='2.2.0+cpu'
1119
export onnx_version='1.15.0'
1220
export onnxruntime_version='1.17.1'
1321
export mxnet_version='1.9.1'
14-
else
15-
export tensorflow_version='2.14.0'
16-
export pytorch_version='2.1.0+cpu'
17-
export torchvision_version='0.16.0+cpu'
18-
export ipex_version='2.1.0+cpu'
19-
export onnx_version='1.14.1'
20-
export onnxruntime_version='1.16.3'
21-
export mxnet_version='1.9.1'
2222
fi
2323

2424

.azure-pipelines/scripts/ut/run_basic_itex.sh

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,10 +4,10 @@ test_case="run basic itex"
44
echo "${test_case}"
55

66
echo "specify fwk version..."
7-
export itex_version='2.14.0.1'
8-
export tensorflow_version='2.14.0-official'
9-
export onnx_version='1.14.1'
10-
export onnxruntime_version='1.16.3'
7+
export itex_version='2.15.0.0'
8+
export tensorflow_version='2.15.0-official'
9+
export onnx_version='1.16.0'
10+
export onnxruntime_version='1.18.0'
1111

1212
echo "set up UT env..."
1313
bash /neural-compressor/.azure-pipelines/scripts/ut/env_setup.sh "${test_case}"

.azure-pipelines/scripts/ut/run_basic_pt_pruning.sh

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -4,9 +4,9 @@ test_case="run basic pt pruning"
44
echo "${test_case}"
55

66
echo "specify fwk version..."
7-
export pytorch_version='2.1.0+cpu'
8-
export torchvision_version='0.16.0+cpu'
9-
export ipex_version='2.1.0+cpu'
7+
export pytorch_version='2.3.0+cpu'
8+
export torchvision_version='0.18.0+cpu'
9+
export ipex_version='2.3.0+cpu'
1010

1111
echo "set up UT env..."
1212
bash /neural-compressor/.azure-pipelines/scripts/ut/env_setup.sh "${test_case}"

.azure-pipelines/template/docker-template.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -33,7 +33,7 @@ steps:
3333
- ${{ if eq(parameters.dockerConfigName, 'commonDockerConfig') }}:
3434
- script: |
3535
rm -fr ${BUILD_SOURCESDIRECTORY} || sudo rm -fr ${BUILD_SOURCESDIRECTORY} || true
36-
echo y | docker system prune
36+
echo y | docker system prune --all
3737
displayName: "Clean workspace"
3838
3939
- checkout: self
@@ -45,7 +45,7 @@ steps:
4545
rm -fr ${BUILD_SOURCESDIRECTORY} || sudo rm -fr ${BUILD_SOURCESDIRECTORY} || true
4646
mkdir ${BUILD_SOURCESDIRECTORY}
4747
chmod 777 ${BUILD_SOURCESDIRECTORY}
48-
echo y | docker system prune
48+
echo y | docker system prune --all
4949
displayName: "Clean workspace"
5050
5151
- checkout: none

README.md

Lines changed: 3 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -195,8 +195,9 @@ quantized_model = fit(model=float_model, conf=static_quant_conf, calib_dataloade
195195
> Further documentations can be found at [User Guide](https://github.com/intel/neural-compressor/blob/master/docs/source/user_guide.md).
196196
197197
## Selected Publications/Events
198-
* Blog by Intel: [Accelerate Meta* Llama 3 with Intel AI Solutions](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-meta-llama3-with-intel-ai-solutions.html)
199-
* Blog by Intel: [Effective Weight-Only Quantization for Large Language Models with Intel® Neural Compressor](https://community.intel.com/t5/Blogs/Tech-Innovation/Artificial-Intelligence-AI/Effective-Weight-Only-Quantization-for-Large-Language-Models/post/1529552) (Oct 2023)
198+
* Blog by Intel: [Neural Compressor: Boosting AI Model Efficiency](https://community.intel.com/t5/Blogs/Tech-Innovation/Artificial-Intelligence-AI/Neural-Compressor-Boosting-AI-Model-Efficiency/post/1604740) (June 2024)
199+
* Blog by Intel: [Optimization of Intel AI Solutions for Alibaba Cloud’s Qwen2 Large Language Models](https://www.intel.com/content/www/us/en/developer/articles/technical/intel-ai-solutions-accelerate-alibaba-qwen2-llms.html) (June 2024)
200+
* Blog by Intel: [Accelerate Meta* Llama 3 with Intel AI Solutions](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-meta-llama3-with-intel-ai-solutions.html) (Apr 2024)
200201
* EMNLP'2023 (Under Review): [TEQ: Trainable Equivalent Transformation for Quantization of LLMs](https://openreview.net/forum?id=iaI8xEINAf&referrer=%5BAuthor%20Console%5D) (Sep 2023)
201202
* arXiv: [Efficient Post-training Quantization with FP8 Formats](https://arxiv.org/abs/2309.14592) (Sep 2023)
202203
* arXiv: [Optimize Weight Rounding via Signed Gradient Descent for the Quantization of LLMs](https://arxiv.org/abs/2309.05516) (Sep 2023)

docker/Dockerfile

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -13,7 +13,7 @@
1313
# See the License for the specific language governing permissions and
1414
# limitations under the License.
1515

16-
ARG UBUNTU_VER=20.04
16+
ARG UBUNTU_VER=22.04
1717
FROM ubuntu:${UBUNTU_VER} as deploy
1818

1919
# See http://bugs.python.org/issue19846

docs/source/publication_list.md

Lines changed: 4 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,6 +1,8 @@
1-
Full Publications/Events (80)
1+
Full Publications/Events (82)
22
==========
3-
## 2024 (1)
3+
## 2024 (3)
4+
* Blog by Intel: [Neural Compressor: Boosting AI Model Efficiency](https://community.intel.com/t5/Blogs/Tech-Innovation/Artificial-Intelligence-AI/Neural-Compressor-Boosting-AI-Model-Efficiency/post/1604740) (June 2024)
5+
* Blog by Intel: [Optimization of Intel AI Solutions for Alibaba Cloud’s Qwen2 Large Language Models](https://www.intel.com/content/www/us/en/developer/articles/technical/intel-ai-solutions-accelerate-alibaba-qwen2-llms.html) (June 2024)
46
* Blog by Intel: [Accelerate Meta* Llama 3 with Intel AI Solutions](https://www.intel.com/content/www/us/en/developer/articles/technical/accelerate-meta-llama3-with-intel-ai-solutions.html) (Apr 2024)
57

68
## 2023 (25)

examples/3.x_api/pytorch/nlp/huggingface_models/language-modeling/quantization/llm/run_clm_no_trainer.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -276,7 +276,7 @@ def get_user_model():
276276
max_seq_length=args.gptq_max_seq_length,
277277
)
278278
dataloader_for_calibration = dataloaderPreprocessor.get_prepared_dataloader()
279-
from neural_compressor.torch.algorithms.weight_only.gptq import move_input_to_device
279+
from neural_compressor.torch.algorithms.weight_only.utility import move_input_to_device
280280
from tqdm import tqdm
281281
def run_fn_for_gptq(model, dataloader_for_calibration, *args):
282282
for batch in tqdm(dataloader_for_calibration):

0 commit comments

Comments
 (0)