Skip to content

Commit d175f44

Browse files
Merge branch 'feature/puzzletron' into jrausch/distillation-consolidation
2 parents 3e39174 + 38d9522 commit d175f44

73 files changed

Lines changed: 373 additions & 190 deletions

File tree

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

examples/puzzletron/evaluation/hf_deployable_anymodel.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -331,7 +331,6 @@ def get_triton_input(self):
331331
Tensor(name="top_p", shape=(-1,), dtype=np.single, optional=True),
332332
Tensor(name="temperature", shape=(-1,), dtype=np.single, optional=True),
333333
Tensor(name="random_seed", shape=(-1,), dtype=np.int_, optional=True),
334-
Tensor(name="max_length", shape=(-1,), dtype=np.int_, optional=True),
335334
Tensor(name="output_logits", shape=(-1,), dtype=np.bool_, optional=True),
336335
Tensor(name="output_scores", shape=(-1,), dtype=np.bool_, optional=True),
337336
)

modelopt/torch/puzzletron/__init__.py

Lines changed: 0 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,6 @@
2727
puzzletron_nas_plugin,
2828
replacement_library,
2929
scoring,
30-
sewing_kit,
3130
subblock_stats,
3231
tools,
3332
utils,

modelopt/torch/puzzletron/activation_scoring/score_pruning_activations.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -22,6 +22,8 @@
2222

2323
from ..tools.logger import mprint
2424

25+
__all__ = ["launch_score_activations"]
26+
2527

2628
def has_checkpoint_support(activation_hooks_kwargs: dict) -> bool:
2729
"""Determine if the activation hook method has proper checkpoint support implemented.

modelopt/torch/puzzletron/anymodel/models/gpt_oss/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -18,5 +18,5 @@
1818

1919
"""GPT-OSS model support for AnyModel."""
2020

21-
from .gpt_oss_converter import GptOssConverter
22-
from .gpt_oss_model_descriptor import GptOssModelDescriptor
21+
from .gpt_oss_converter import *
22+
from .gpt_oss_model_descriptor import *

modelopt/torch/puzzletron/anymodel/models/gpt_oss/gpt_oss_converter.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,8 @@
2323
from ....block_config import AttentionConfig, BlockConfig, FFNConfig, MoEConfig
2424
from ...converter import Converter, ConverterFactory
2525

26+
__all__ = ["GptOssConverter"]
27+
2628

2729
@ConverterFactory.register_decorator("gpt_oss")
2830
class GptOssConverter(Converter):

modelopt/torch/puzzletron/anymodel/models/gpt_oss/gpt_oss_model_descriptor.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -37,6 +37,8 @@
3737
from ...model_descriptor import ModelDescriptor, ModelDescriptorFactory
3838
from ...puzzformer.no_op import MatchingZeros, Same, return_tuple_of_size
3939

40+
__all__ = ["GptOssModelDescriptor", "GptOssExpertRemovalLayerDescriptor"]
41+
4042

4143
@ModelDescriptorFactory.register_decorator("gpt_oss")
4244
class GptOssModelDescriptor(ModelDescriptor):

modelopt/torch/puzzletron/anymodel/models/gpt_oss/gpt_oss_pruned_to_mxfp4.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -35,6 +35,8 @@
3535
from tqdm import tqdm
3636
from transformers.integrations.mxfp4 import convert_moe_packed_tensors
3737

38+
__all__ = []
39+
3840

3941
def deduce_experts_for_layer(
4042
layer: int,

modelopt/torch/puzzletron/anymodel/models/llama/__init__.py

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -13,5 +13,5 @@
1313
# See the License for the specific language governing permissions and
1414
# limitations under the License.
1515

16-
from .llama_converter import LlamaConverter
17-
from .llama_model_descriptor import LlamaModelDescriptor
16+
from .llama_converter import *
17+
from .llama_model_descriptor import *

modelopt/torch/puzzletron/anymodel/models/llama/llama_converter.py

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -23,6 +23,8 @@
2323
from ....block_config import AttentionConfig, BlockConfig, FFNConfig
2424
from ...converter import Converter, ConverterFactory
2525

26+
__all__ = ["LlamaConverter"]
27+
2628

2729
@ConverterFactory.register_decorator("llama")
2830
class LlamaConverter(Converter):

modelopt/torch/puzzletron/anymodel/models/llama/llama_model_descriptor.py

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -32,6 +32,12 @@
3232
from ...model_descriptor import ModelDescriptor, ModelDescriptorFactory
3333
from ...puzzformer.no_op import MatchingZeros, Same, return_tuple_of_size
3434

35+
__all__ = [
36+
"LlamaModelDescriptor",
37+
"LlamaFFNIntermediateLayerDescriptor",
38+
"LlamaKVHeadsLayerDescriptor",
39+
]
40+
3541

3642
@ModelDescriptorFactory.register_decorator("llama")
3743
class LlamaModelDescriptor(ModelDescriptor):

0 commit comments

Comments
 (0)