Skip to content

float_quantize at multi-gpu works wrong. #67

@jinsol-neubla

Description

@jinsol-neubla

When I run model after applying float_quantize to weight or activation with multi-GPU,
(huggingface opt-model with device_map='auto')
quantization of layers allocated to second or later gpu works wrong.
The output of quantization shows mostly 0-value.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions