Skip to content
Merged
Changes from 1 commit
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
use safe threading layer
Signed-off-by: Kaihui-intel <[email protected]>
  • Loading branch information
Kaihui-intel committed Jul 26, 2024
commit eb5f7912d830ff1f27d325a296d7c541362b9e82
1 change: 1 addition & 0 deletions neural_compressor/torch/algorithms/weight_only/modules.py
Original file line number Diff line number Diff line change
Expand Up @@ -607,6 +607,7 @@ def pack_array_with_numba(

pack_method_name = f"pack_array_with_numba_b{bits}_c{compress_bits}"
pack_method = getattr(self, pack_method_name)
numba.config.THREADING_LAYER = "safe"
return pack_method(raw_array, packed_array, n_pack, new_in_features)

def pack_tensor_with_numpy_impl(self, raw_tensor):
Expand Down