Skip to content

Commit

Permalink
add gh issue to comment
Browse files Browse the repository at this point in the history
  • Loading branch information
samuel100 committed Dec 31, 2024
1 parent 4218e36 commit 307c960
Showing 1 changed file with 2 additions and 1 deletion.
3 changes: 2 additions & 1 deletion olive/passes/pytorch/gptq.py
Original file line number Diff line number Diff line change
Expand Up @@ -160,7 +160,8 @@ def _run_for_config(
quantized_model: BaseGPTQForCausalLM = model_class(pytorch_model, False, quantize_config)

# explicitly move quantized model to CUDA device to avoid the "Expected all tensors to be
# on the same device" error in auto-gptq
# on the same device" error in auto-gptq.
# see https://github.com/AutoGPTQ/AutoGPTQ/issues/729
quantized_model.to("cuda")

fields_to_set = {
Expand Down

0 comments on commit 307c960

Please sign in to comment.