-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Description
微调的数据集图片一定要处理成统一尺寸吗?新做的数据集一直报错
Traceback (most recent call last):
File "/home/partners/ms-swift/swift/cli/sft.py", line 10, in
sft_main()
File "/home/partners/ms-swift/swift/llm/train/sft.py", line 341, in sft_main
return SwiftSft(args).main()
File "/home/partners/ms-swift/swift/llm/base.py", line 49, in main
result = self.run()
File "/home/partners/ms-swift/swift/llm/train/sft.py", line 196, in run
return self.train(trainer)
File "/home/partners/ms-swift/swift/llm/train/sft.py", line 244, in train
trainer.train(trainer.args.resume_from_checkpoint)
File "/home/partners/ms-swift/swift/trainers/mixin.py", line 801, in train
res = super().train(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/transformers/trainer.py", line 2123, in train
return inner_training_loop(
File "/home/partners/swift_env/lib/python3.10/site-packages/transformers/trainer.py", line 2481, in _inner_training_loop
tr_loss_step = self.training_step(model, inputs, num_items_in_batch)
File "/home/partners/ms-swift/swift/trainers/trainers.py", line 409, in training_step
return super().training_step(model, inputs, *args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/transformers/trainer.py", line 3579, in training_step
loss = self.compute_loss(model, inputs, num_items_in_batch=num_items_in_batch)
File "/home/partners/ms-swift/swift/trainers/trainers.py", line 323, in compute_loss
outputs = model(**inputs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1881, in _call_impl
return inner()
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1829, in inner
result = forward_call(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/accelerate/utils/operations.py", line 819, in forward
return model_forward(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/accelerate/utils/operations.py", line 807, in call
return convert_to_fp32(self.model_forward(*args, **kwargs))
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/amp/autocast_mode.py", line 44, in decorate_autocast
return func(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward
return self.base_model(
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward
return self.model.forward(*args, **kwargs)
File "/home/partners/.cache/huggingface/modules/transformers_modules/DeepSeek-OCR/modeling_deepseekocr.py", line 565, in forward
outputs = self.model(
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/partners/.cache/huggingface/modules/transformers_modules/DeepSeek-OCR/modeling_deepseekocr.py", line 448, in forward
global_features = global_features.view(h, w, n_dim)
RuntimeError: shape '[16, 16, 1280]' is invalid for input of size 655360