Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

How #8

Open
wuzeww opened this issue Oct 25, 2021 · 29 comments
Open

How #8

wuzeww opened this issue Oct 25, 2021 · 29 comments
Assignees

Comments

@wuzeww
Copy link

wuzeww commented Oct 25, 2021

Are giga, vgn and giga-aff are the same epoches? Are they all 20 epoches?

@wuzeww wuzeww closed this as completed Oct 25, 2021
@wuzeww wuzeww reopened this Oct 25, 2021
@Steve-Tod
Copy link
Collaborator

Yeah, I think so.

@wuzeww
Copy link
Author

wuzeww commented Oct 26, 2021

But the result of GIGA-Aff is higher than the paper.

@wuzeww
Copy link
Author

wuzeww commented Oct 26, 2021

I am currently making a graduation project based on your thesis, so I am eager to know the specific epochs of your work. You can also reply me by email [email protected].
Extremely grateful!

@Steve-Tod
Copy link
Collaborator

But the result of GIGA-Aff is higher than the paper.

That's possible, different devices and different random seeds can give different results. I suggest running more tests with more different random seeds. BTW, how much higher?

@wuzeww
Copy link
Author

wuzeww commented Oct 26, 2021

5 or 6 percentage points

@Steve-Tod
Copy link
Collaborator

Is that the average result from multiple random seeds?

@wuzeww
Copy link
Author

wuzeww commented Oct 26, 2021

Yes, the random seeds are [0, 1, 2, 3, 4]. The epoches are 20. I tested twice, the results were both higher than the paper.

@Steve-Tod
Copy link
Collaborator

Hmmm, how about the result of GIGA? Is it better than GIGA-Aff?

@wuzeww
Copy link
Author

wuzeww commented Oct 27, 2021 via email

@Steve-Tod
Copy link
Collaborator

Hmmm, that's weird. Have you checked the loss curve and made sure they both converged?

@wuzeww
Copy link
Author

wuzeww commented Oct 28, 2021

This is the loss graph of GIGA-Aff training 20 epoches

2021-10-28 09-59-19 的屏幕截图

This is the 10 epoches

2021-10-28 09-59-26 的屏幕截图

@wuzeww
Copy link
Author

wuzeww commented Oct 29, 2021

Hi,

I trained GIGA-Aff for 10 epoches, the result is higher 5 percentage points than the paper. I wonder if there is a problem with the code.

@Steve-Tod
Copy link
Collaborator

How about the training figure of GIGA? Have you trained GIGA?

@wuzeww
Copy link
Author

wuzeww commented Oct 30, 2021

This the loss curve of giga:
2021-10-30 20-51-43 的屏幕截图

@Steve-Tod
Copy link
Collaborator

So the GIGA trained with the same number of epochs perform worse than GIGA-Aff?

@wuzeww
Copy link
Author

wuzeww commented Oct 31, 2021 via email

@Steve-Tod
Copy link
Collaborator

Hmmm, that's weird. What scenario are you using? Packed or pile?

@wuzeww
Copy link
Author

wuzeww commented Nov 1, 2021 via email

@Steve-Tod
Copy link
Collaborator

Not sure about that. GIGA should perform better than GIGA-Aff, especially in packed scenarios.

@wuzeww
Copy link
Author

wuzeww commented Nov 3, 2021

Sorry to reply you now.

It's weird. I retrained giga-aff, and the result was lower than before, even lower than the paper.

@Steve-Tod
Copy link
Collaborator

Do you retrain with the same setting?

@wuzeww
Copy link
Author

wuzeww commented Nov 3, 2021 via email

@Steve-Tod
Copy link
Collaborator

The absolute value may vary because the test scenes can be different, but the relative performance between GIGA and GIGA-Aff should stay the same. However, you said GIGA is worse than GIGA-Aff previously, which is very weird. Did you train GIGA and GIGA-Aff on the same computer?

@wuzeww
Copy link
Author

wuzeww commented Nov 5, 2021

The absolute value may vary because the test scenes can be different, but the relative performance between GIGA and GIGA-Aff should stay the same. However, you said GIGA is worse than GIGA-Aff previously, which is very weird. Did you train GIGA and GIGA-Aff on the same computer?

Yes, I trained them on the same computer before. Training the same model on the same computer, the loss curve obtained is different.

1 similar comment
@wuzeww
Copy link
Author

wuzeww commented Nov 5, 2021

The absolute value may vary because the test scenes can be different, but the relative performance between GIGA and GIGA-Aff should stay the same. However, you said GIGA is worse than GIGA-Aff previously, which is very weird. Did you train GIGA and GIGA-Aff on the same computer?

Yes, I trained them on the same computer before. Training the same model on the same computer, the loss curve obtained is different.

@Steve-Tod
Copy link
Collaborator

The loss curve can be different. I think training on different computers is OK. The important thing is testing on the same computer so that after fixing the random seed, the generated scenes will be the same. (I should have asked if you test them on the same computer, it was a typo.)

@wuzeww
Copy link
Author

wuzeww commented Nov 5, 2021 via email

@Steve-Tod
Copy link
Collaborator

Not sure why this happens. I'll look back to this and re-train by myself later.

@Steve-Tod Steve-Tod self-assigned this Nov 5, 2021
@Steve-Tod Steve-Tod pinned this issue Nov 5, 2021
@Steve-Tod Steve-Tod unpinned this issue Nov 5, 2021
@wuzeww
Copy link
Author

wuzeww commented Nov 5, 2021 via email

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants