Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion fms_mo/custom_ext_kernels/triton_kernels.py
Original file line number Diff line number Diff line change
Expand Up @@ -449,7 +449,7 @@ def round_and_trun_mask(chunk_trun_bits, clamp_acc_to_dl16):
@triton.jit
def round_and_trun(x, round_bit, trun_mask, clamp_acc_to_dl16):
"""Round and truncate (usually for accumulator)."""
x = libdevice.uint_as_float((libdevice.float_as_uint(x) + round_bit) & trun_mask)
x = libdevice.uint_as_float((libdevice.float_as_uint(x) + round_bit) & trun_mask) # pylint: disable=assignment-from-no-return

if clamp_acc_to_dl16:
# clamp to DL16 min/max:
Expand Down
1 change: 0 additions & 1 deletion fms_mo/run_quant.py
Original file line number Diff line number Diff line change
Expand Up @@ -155,7 +155,6 @@ def run_gptq(model_args, data_args, opt_args, gptq_args):
v2_memory_device="cpu",
)


# Add custom model_type mapping to gptqmodel LUT so GPTQModel can recognize them.
for mtype, cls in custom_gptq_classes.items():
if mtype in MODEL_MAP:
Expand Down
5 changes: 2 additions & 3 deletions fms_mo/training_args.py
Original file line number Diff line number Diff line change
Expand Up @@ -139,8 +139,8 @@ class OptArguments(TypeChecker):

quant_method: str = field(
metadata={
"choices": ["gptq", "gptqv2", "fp8", "dq"],
"help": "Quantization technique"
"choices": ["gptq", "gptqv2", "fp8", "dq"],
"help": "Quantization technique",
}
)
output_dir: str = field(
Expand Down Expand Up @@ -229,7 +229,6 @@ class GPTQArguments(TypeChecker):
cache_examples_on_gpu: bool = True



@dataclass
class FP8Arguments(TypeChecker):
"""Dataclass for FP8 related arguments that will be used by llm-compressor."""
Expand Down
Loading