Skip to content

Commit 07de89f

Browse files
authored
fix phi3-vision bug (#1039)
1 parent 05b4002 commit 07de89f

File tree

2 files changed

+8
-4
lines changed

2 files changed

+8
-4
lines changed

swift/llm/sft.py

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -151,9 +151,7 @@ def llm_sft(args: SftArguments) -> Dict[str, Union[str, Any]]:
151151
if args.gradient_checkpointing:
152152
model.config.use_cache = False # fix transformers==4.36
153153
logger.info('Setting model.config.use_cache: False')
154-
disable_require_grads = MODEL_MAPPING[args.model_type].get('disable_require_grads', False)
155-
if not disable_require_grads:
156-
model.enable_input_require_grads()
154+
model.enable_input_require_grads()
157155

158156
if use_torchacc():
159157
model.config.use_cache = False

swift/llm/utils/model.py

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1160,14 +1160,17 @@ def get_model_tokenizer_paligemma_vision(model_dir: str,
11601160
return model, tokenizer
11611161

11621162

1163+
def _clone_grads(module, input, output):
1164+
return output.requires_grad_(True).clone()
1165+
1166+
11631167
@register_model(
11641168
ModelType.phi3_vision_128k_instruct,
11651169
'LLM-Research/Phi-3-vision-128k-instruct',
11661170
LoRATM.phi3,
11671171
TemplateType.phi3_vl,
11681172
support_flash_attn=True,
11691173
requires=['transformers>=4.36'],
1170-
disable_require_grads=True,
11711174
tags=['multi-modal', 'vision'],
11721175
hf_model_id='microsoft/Phi-3-vision-128k-instruct')
11731176
def get_model_tokenizer_phi3_vision(model_dir: str,
@@ -1180,6 +1183,9 @@ def get_model_tokenizer_phi3_vision(model_dir: str,
11801183
model, tokenizer = get_model_tokenizer_with_flash_attn(model_dir, torch_dtype, model_kwargs, load_model, **kwargs)
11811184
tokenizer.processor = processor
11821185

1186+
if load_model:
1187+
model.model.vision_embed_tokens.wte.register_forward_hook(_clone_grads)
1188+
11831189
return model, tokenizer
11841190

11851191

0 commit comments

Comments
 (0)