Skip to content

微调数据集 #233

@hharch

Description

@hharch

微调的数据集图片一定要处理成统一尺寸吗?新做的数据集一直报错
Traceback (most recent call last):
File "/home/partners/ms-swift/swift/cli/sft.py", line 10, in
sft_main()
File "/home/partners/ms-swift/swift/llm/train/sft.py", line 341, in sft_main
return SwiftSft(args).main()
File "/home/partners/ms-swift/swift/llm/base.py", line 49, in main
result = self.run()
File "/home/partners/ms-swift/swift/llm/train/sft.py", line 196, in run
return self.train(trainer)
File "/home/partners/ms-swift/swift/llm/train/sft.py", line 244, in train
trainer.train(trainer.args.resume_from_checkpoint)
File "/home/partners/ms-swift/swift/trainers/mixin.py", line 801, in train
res = super().train(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/transformers/trainer.py", line 2123, in train
return inner_training_loop(
File "/home/partners/swift_env/lib/python3.10/site-packages/transformers/trainer.py", line 2481, in _inner_training_loop
tr_loss_step = self.training_step(model, inputs, num_items_in_batch)
File "/home/partners/ms-swift/swift/trainers/trainers.py", line 409, in training_step
return super().training_step(model, inputs, *args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/transformers/trainer.py", line 3579, in training_step
loss = self.compute_loss(model, inputs, num_items_in_batch=num_items_in_batch)
File "/home/partners/ms-swift/swift/trainers/trainers.py", line 323, in compute_loss
outputs = model(**inputs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1881, in _call_impl
return inner()
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1829, in inner
result = forward_call(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/accelerate/utils/operations.py", line 819, in forward
return model_forward(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/accelerate/utils/operations.py", line 807, in call
return convert_to_fp32(self.model_forward(*args, **kwargs))
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/amp/autocast_mode.py", line 44, in decorate_autocast
return func(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/peft/peft_model.py", line 1850, in forward
return self.base_model(
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/peft/tuners/tuners_utils.py", line 222, in forward
return self.model.forward(*args, **kwargs)
File "/home/partners/.cache/huggingface/modules/transformers_modules/DeepSeek-OCR/modeling_deepseekocr.py", line 565, in forward
outputs = self.model(
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1775, in _wrapped_call_impl
return self._call_impl(*args, **kwargs)
File "/home/partners/swift_env/lib/python3.10/site-packages/torch/nn/modules/module.py", line 1786, in _call_impl
return forward_call(*args, **kwargs)
File "/home/partners/.cache/huggingface/modules/transformers_modules/DeepSeek-OCR/modeling_deepseekocr.py", line 448, in forward
global_features = global_features.view(h, w, n_dim)
RuntimeError: shape '[16, 16, 1280]' is invalid for input of size 655360

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions