Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

loss sometimes jumpes to 200 #11

Open
yzho0907 opened this issue Apr 18, 2018 · 2 comments
Open

loss sometimes jumpes to 200 #11

yzho0907 opened this issue Apr 18, 2018 · 2 comments

Comments

@yzho0907
Copy link

my params are {
"cell_type": "lstm",
"depth": 2,
"attention_type": "Luong",
"bidirectional": true,
"use_residual": true,
"use_dropout": false,
"time_major": true,
"hidden_units": 1024,
"optimizer": "adam",
"learning_rate": 0.001
}
and i am using fasttext pre-trained word2vec as well
batch size is 128 which i haven't changed it.
But the loss is always fluctuating and sometimes jumps to a huge number (200), it usually started as 10-20. It's cuz this version of seq2seq add a part of reinforcement learning? plz tell u how to fix it?

@yzho0907
Copy link
Author

我把词向量去掉,把参数调到默认,loss有时候也会跳到一个很大的数字,请问下怎么解决啊?训练出来的对话也完全不对。

@yzho0907
Copy link
Author

训练轮数增加后loss就会变的很巨大,请问有什么办法可以解决嘛?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant