Skip to content

Commit

Permalink
add known issue of EmbeddingBag INT8 accuracy loss (#1017)
Browse files Browse the repository at this point in the history
* add known issue

restate the embedding issue

* add release notes for 1.12.100
  • Loading branch information
jingxu10 authored Aug 4, 2022
1 parent f092200 commit e2fbae0
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 0 deletions.
2 changes: 2 additions & 0 deletions docs/tutorials/performance_tuning/known_issues.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,8 @@
Known Issues
============

- Supporting of EmbeddingBag with INT8 when bag size > 1 is working in progress.

- Compiling with gcc 11 might result in `illegal instruction` error.

- `RuntimeError: Overflow when unpacking long` when a tensor's min max value exceeds int range while performing int8 calibration. Please customize QConfig to use min-max calibration method.
Expand Down
4 changes: 4 additions & 0 deletions docs/tutorials/releases.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,10 @@
Releases
=============

## 1.12.100

This is a patch release to fix the AVX2 issue that blocks running on non-AVX512 platforms.

## 1.12.0

We are excited to bring you the release of Intel® Extension for PyTorch\* 1.12.0-cpu, by tightly following PyTorch [1.12](https://github.com/pytorch/pytorch/releases/tag/v1.12.0) release. In this release, we matured the automatic int8 quantization and made it a stable feature. We stabilized runtime extension and brought about a MultiStreamModule feature to further boost throughput in offline inference scenario. We also brought about various enhancements in operation and graph which are positive for performance of broad set of workloads.
Expand Down

0 comments on commit e2fbae0

Please sign in to comment.