k-l-lambda commited on
Commit
9feeab3
·
1 Parent(s): f9ec70e

apilog_k25_mtp/output_v3/checkpoints/7

Browse files
Files changed (3) hide show
  1. .gitignore +4 -0
  2. modeling_deepseek.py +4 -1
  3. mtp.safetensors +3 -0
.gitignore ADDED
@@ -0,0 +1,4 @@
 
 
 
 
 
1
+ *.safetensors
2
+ *.pyc
3
+
4
+ !mtp.safetensors
modeling_deepseek.py CHANGED
@@ -46,7 +46,10 @@ from transformers.utils import (add_start_docstrings,
46
  replace_return_docstrings)
47
  from transformers.utils.import_utils import is_torch_fx_available
48
 
49
- from .configuration_deepseek import DeepseekV3Config
 
 
 
50
 
51
  if is_flash_attn_2_available():
52
  from flash_attn import flash_attn_func, flash_attn_varlen_func
 
46
  replace_return_docstrings)
47
  from transformers.utils.import_utils import is_torch_fx_available
48
 
49
+ try:
50
+ from .configuration_deepseek import DeepseekV3Config
51
+ except ImportError:
52
+ from configuration_deepseek import DeepseekV3Config
53
 
54
  if is_flash_attn_2_available():
55
  from flash_attn import flash_attn_func, flash_attn_varlen_func
mtp.safetensors ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0765cc114a73b725b3c08b88229f877956629a8111c7268cef879d846d3c2abc
3
+ size 14712186504