Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

cross-scale problem #1

Open
chchshshhh opened this issue Sep 25, 2022 · 2 comments
Open

cross-scale problem #1

chchshshhh opened this issue Sep 25, 2022 · 2 comments

Comments

@chchshshhh
Copy link

Very interesting work, but why only the loss of two scales is crossed, and A1A2 is not added for cross-scale loss?

@TheoPis
Copy link
Collaborator

TheoPis commented Sep 26, 2022

Thanks for the question. In general, there is no reason not to add more cross-scale terms (e.g A1-A2 A2-A3 ...). Adding terms would slightly increase training time but it is possible it could improve performance. In practice, we opted for cross-scale terms with features that have a slightly higher scale difference than A1-A2 (x2), namely A1-A3 (x4) and A1-A4 (x8). The motivation here was to better capture local vs global properties of the classes in the contrastive loss. We have not systematically tested all other possible cross-scale term options as we found that A1-A3 and A1-A4 worked well for many models/datasets.

@chchshshhh
Copy link
Author

Thanks for the question. In general, there is no reason not to add more cross-scale terms (e.g A1-A2 A2-A3 ...). Adding terms would slightly increase training time but it is possible it could improve performance. In practice, we opted for cross-scale terms with features that have a slightly higher scale difference than A1-A2 (x2), namely A1-A3 (x4) and A1-A4 (x8). The motivation here was to better capture local vs global properties of the classes in the contrastive loss. We have not systematically tested all other possible cross-scale term options as we found that A1-A3 and A1-A4 worked well for many models/datasets.

Thank you very much for your answer. I have another question about the weight of loss, why does the setting of [1.0 0.7 0.4 0.1] work well? Is there any basis for that? or do I understand that in cross-scale learning, shallow contrast learning is more important and can lead to deeper semantic feature learning?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants