Skip to content

Commit

Permalink
fix deepseek_vl2
Browse files Browse the repository at this point in the history
  • Loading branch information
Jintao-Huang committed Feb 23, 2025
1 parent 2433f48 commit df4bc70
Show file tree
Hide file tree
Showing 2 changed files with 1 addition and 3 deletions.
1 change: 0 additions & 1 deletion swift/llm/model/model/deepseek.py
Original file line number Diff line number Diff line change
Expand Up @@ -216,7 +216,6 @@ def get_model_tokenizer_deepseek_vl2(model_dir: str, *args, **kwargs):
except ImportError:
# compat transformers>=4.42
import transformers
transformers.models.llama.modeling_llama.LlamaAttention = None
transformers.models.llama.modeling_llama.LlamaFlashAttention2 = None
from deepseek_vl2.models import DeepseekVLV2Processor
processor: DeepseekVLV2Processor = DeepseekVLV2Processor.from_pretrained(model_dir)
Expand Down
3 changes: 1 addition & 2 deletions swift/llm/template/template/deepseek.py
Original file line number Diff line number Diff line change
Expand Up @@ -185,9 +185,8 @@ def generate(self, model, *args, **kwargs):
return {'sequences': generated_tokens}

def decode(self, generate_ids: List[int], is_finished: bool = True, tokenizer_kwargs=None, **kwargs) -> Any:
if not kwargs['template_inputs'].generate_mode:
if 'template_inputs' not in kwargs or not kwargs['template_inputs'].generate_mode:
return super().decode(generate_ids, is_finished, tokenizer_kwargs, **kwargs)

else:
img_size = get_env_args('img_size', int, 384)
patch_size = 16
Expand Down

0 comments on commit df4bc70

Please sign in to comment.