Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Question] Abnormal Generation after finetuning ONLY single label classification dataset #1796

Open
enkaranfiles opened this issue Dec 11, 2024 · 0 comments

Comments

@enkaranfiles
Copy link

enkaranfiles commented Dec 11, 2024

Question

I am trying to change the model architecture,

Model exhibits abnormal generation behavior, such as repetitive or nonsensical outputs, despite successful pretraining performance.

During pretraining, the model performed well on scene description tasks and demonstrated strong zero-shot classification capabilities, indicating effective alignment between the visual encoder and the LLM.

Example generation after pretraining:

  • I've identified the one applicable category for this image. The category is: Arable land.

At that stage, model pretrained only scene description and do not know anything about the scenes class names.

However, after fine-tuning on the classification dataset, the model fails to produce meaningful outputs.

Example generation of the after finetuning on single label classification dataset:

-xResSeawayResSeaidentialSeaidentialSeaSeaSeawaySeaSeaSeawaySeaSeawayResSeaSeaSeaidentialSeaSeaSeaSeaSeaSeaidentialSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSeaSea

So in the high-level, I am trying change vision encoder for processing different domain input image (dataset curated by me), and in the finetuning stage I am using single label classification dataset. (followed LLaVA conservation template)

I think getting good result after pretraining stage indicates, we are getting succesful signal after pretraining, means nothing wrong with vision_tower section. I am thinking possible reason for this issue: Pretrained task and finetuning task mis-alignment.

Should I increase the diffuculty level of the instruction tuning dataset?

PS: There were some other issues related to abnormal and repetetive generation, so I checked the image broker and dataset folder. They were all fine.

PS: Loss function has gradually decrease over the steps.

PS: I have tried to model overfitting on the given train set by decreasing sample size(trained only 3 percent of the finetuning data) and increasing number of epoch. But still cannot perform well on train set.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant