|
38 | 38 | from modelopt.torch.utils import create_param_grad_clear_hook, print_rank_0, report_memory |
39 | 39 | from modelopt.torch.utils.distributed import DistributedProcessGroup, ParallelState, is_master |
40 | 40 |
|
41 | | -from . import config as mtq_config |
42 | | -from . import model_calib |
43 | | -from .config import QuantizeConfig, QuantizerAttributeConfig |
44 | | -from .conversion import set_quantizer_by_cfg |
45 | | -from .nn import QuantLinearConvBase, QuantModule, SequentialQuantizer, TensorQuantizer |
46 | | -from .utils import is_quantized_linear |
| 41 | +from .. import config as mtq_config |
| 42 | +from .. import model_calib |
| 43 | +from ..config import QuantizeConfig, QuantizerAttributeConfig |
| 44 | +from ..conversion import set_quantizer_by_cfg |
| 45 | +from ..nn import QuantLinearConvBase, QuantModule, SequentialQuantizer, TensorQuantizer |
| 46 | +from ..utils import is_quantized_linear |
47 | 47 |
|
48 | 48 |
|
49 | 49 | def estimate_quant_compression(quant_cfg: QuantizeConfig) -> float: |
@@ -615,8 +615,8 @@ def before_search(self): |
615 | 615 | # Import here to avoid circular import |
616 | 616 | from modelopt.torch.quantization.model_quant import calibrate |
617 | 617 |
|
618 | | - from .conversion import restore_quantizer_state, update_quantize_metadata |
619 | | - from .utils import get_quantizer_state_dict, set_quantizer_state_dict |
| 618 | + from ..conversion import restore_quantizer_state, update_quantize_metadata |
| 619 | + from ..utils import get_quantizer_state_dict, set_quantizer_state_dict |
620 | 620 |
|
621 | 621 | super().before_search() |
622 | 622 | restored_method = getattr(self, "method", None) |
|
0 commit comments