Skip to content

cannot import name 'check_set_window_size' from 'transformer_engine.pytorch.attention' #1891

@njw1123

Description

@njw1123

transformer_engine=2.4.0

  if depth == 0 or not hasattr(obj, "__dict__") or not isinstance(obj, types.ModuleType):
Traceback (most recent call last):
  File "/jizhicfs/marvinhjia/njw1123/megatron_ori/pretrain_gpt.py", line 12, in <module>
    from megatron.training import get_args
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/training.py", line 27, in <module>
    from megatron.utils import lazy_gc, average_losses_across_data_parallel_group, get_all_moe_loss
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/utils.py", line 37, in <module>
    from megatron.model.module import param_is_not_shared
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/model/__init__.py", line 21, in <module>
    from .bert_model import BertModel
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/model/bert_model.py", line 23, in <module>
    from megatron.model.language_model import parallel_lm_logits
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/model/language_model.py", line 26, in <module>
    from megatron.model.transformer import ParallelTransformer
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/model/transformer.py", line 35, in <module>
    from megatron.model.mamba_utils import allocate_layers, mamba_select_layers_for_pipeline_parallel, _init_weights 
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/model/mamba_utils.py", line 17, in <module>
    from megatron.custom_layers.megatron_layers import TENorm
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/custom_layers/__init__.py", line 2, in <module>
    from .custom_transformer import CustomTransformerLayer, MOETransformerLayer
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/custom_layers/custom_transformer.py", line 18, in <module>
    from .custom_attention import CustomMultiheadAttention, DeepSeekV2Attention
  File "/usr/local/python3/lib/python3.10/site-packages/megatron_lm-1.1.5-py3.10.egg/megatron/custom_layers/custom_attention.py", line 12, in <module>
    from transformer_engine.pytorch.attention import check_set_window_size, _SplitAlongDim, InferenceParams, apply_rotary_pos_emb, DotProductAttention
ImportError: cannot import name 'check_set_window_size' from 'transformer_engine.pytorch.attention' (/usr/local/python3/lib/python3.10/site-packages/transformer_engine/pytorch/attention/__init__.py)

Metadata

Metadata

Assignees

No one assigned

    Labels

    bugSomething isn't working

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions