Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Unit 2: Explaining the "residual learning" #342

Open
0xD4rky opened this issue Sep 5, 2024 · 4 comments
Open

Unit 2: Explaining the "residual learning" #342

0xD4rky opened this issue Sep 5, 2024 · 4 comments

Comments

@0xD4rky
Copy link
Contributor

0xD4rky commented Sep 5, 2024

I would like to explain the residual learning, introduced in the official paper, in depth.

I want to explain how learning (h(x)-x) is easier for the model rather than learning h(x) (where h(x) is the function that maps the input and output of the stacked layer).

Hence, allow me to raise a PR for updating the docs and you review the changes!

@johko
Copy link
Owner

johko commented Sep 19, 2024

Sounds great, feel free to write something up and create a Pr 👍

@0xD4rky
Copy link
Contributor Author

0xD4rky commented Sep 20, 2024

will do for sure!

@sezan92
Copy link
Contributor

sezan92 commented Oct 25, 2024

is this issue done ?

@johko
Copy link
Owner

johko commented Oct 26, 2024

I think it is connected to PR #347, which is still open, but almost done

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

3 participants