Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Some questions about paper and code #10

Open
jsh-hit opened this issue Oct 29, 2024 · 0 comments
Open

Some questions about paper and code #10

jsh-hit opened this issue Oct 29, 2024 · 0 comments

Comments

@jsh-hit
Copy link

jsh-hit commented Oct 29, 2024

Dear authors,

First of all, thank you very much for making your code available and for the interesting work presented in your paper "AdaProp: Learning Adaptive Propagation for Graph Neural Network based Knowledge Graph Reasoning". I have a few questions regarding the implementation and some specific details of the model. I would be grateful if you could provide some clarification on the following points:

I also want to mention that my background in this area is still developing, so some of my questions might seem a bit basic. I hope you don't mind, and I truly appreciate your patience in helping me understand these concepts better.

Inverse Relationships: Regarding the addition of inverse relations, could you please explain the purpose of adding these inverse relations? Does this imply that the directional nature of relationships in the knowledge graph becomes less significant?

Loss Function: It seems that the loss function described in the paper does not exactly match the one implemented in the code. Specifically, I found the following:

loss = torch.sum(- pos_scores + max_n + torch.log(torch.sum(torch.exp(scores - max_n), 1)))

Could you provide some insights into this discrepancy?

Full Propagation Mechanism in CompGCN: In the case of CompGCN, do you consider "Full propagation" to mean that after each GNN layer, the embeddings of all nodes in the graph are updated simultaneously?

Gumbel Sampling: I understand that Gumbel sampling allows for gradient calculation on discrete variables during backpropagation. Could you please elaborate on why you chose to still use the straight-through (ST) estimator in this context?

Thank you in advance for your time and consideration. Your guidance on these questions would be immensely helpful in understanding the details of your implementation.

Best regards.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant