Skip to content

微调代码有bug #19

@sophieyl820

Description

@sophieyl820

System Info / 系統信息

transformers 4.48.0

Who can help? / 谁可以帮助到您?

No response

Information / 问题信息

  • The official example scripts / 官方的示例脚本
  • My own modified scripts / 我自己修改的脚本和任务

Reproduction / 复现过程

我使用微调后的代码输出的结果是
{'posture': 'walking', 'orientation': 'front'}ohlments of privacy and security for all individuals while maintaining their dignity and autonomy.
事实上我只希望到{'posture': 'walking', 'orientation': 'front'}后面的部分不要;

我目前发现微调代码的bug在finetune_vision.py 中的process_batch_eval中,计算eval的loss时label是空会报错,代码这边我修改的是这样

# Left Padding
            padding_length = max(0, max_input_length - len(input_segment))
            input_segment = [tokenizer.pad_token_id] * padding_length + input_segment[:max_input_length]
            attention_segment = [0] * padding_length + attention_segment[:max_input_length]
            position_segment = [0] * padding_length + position_segment[:max_input_length]
            # output_segment = [tokenizer.pad_token_id] * padding_length + output_segment[:max_output_length]
            
            # Right padding for output (target)
            output_segment = output_segment[:max_output_length]
            output_padding = max(0, max_output_length - len(output_segment))
            output_segment = output_segment + [tokenizer.pad_token_id] * output_padding

            batched_input_ids.append(input_segment[:max_input_length])
            batched_attention_mask.append(attention_segment[:max_input_length])
            batched_position_ids.append(position_segment[:max_input_length])
            batched_output_ids.append(output_segment[:max_output_length])

但我总觉得你们的训练微调代码多少也有问题,不然为啥学不到stop word的位置呢?

Expected behavior / 期待表现

求解决

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions