Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

CE(pseudo hard label, teacher_us) #21

Open
miramirakim227 opened this issue Aug 15, 2021 · 1 comment
Open

CE(pseudo hard label, teacher_us) #21

miramirakim227 opened this issue Aug 15, 2021 · 1 comment

Comments

@miramirakim227
Copy link

miramirakim227 commented Aug 15, 2021

image

I have a small question about the above code.

The aim of the cross entropy is to make the two different distributions be similar with each other.
I understood the first CE loss but not the second CE loss, which confuses me a lot.
In line 218, the latter argument of CE should follow the former argument.
In this case is the purpose of CE loss to make the raw output of teacher model to be sharpened?

@Xiao0728
Copy link

Xiao0728 commented Dec 30, 2021

did you figure it out? and how was your implementation results? does it work? Thanks.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants