-
Notifications
You must be signed in to change notification settings - Fork 118
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Attention sparse embeddings #235
Conversation
"History is merely a list of surprises... It can only prepare us to be surprised yet again.", | ||
])) | ||
|
||
for result in output: |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I bet this test could be better :D
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I am open to suggestions
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I would check some types / shapes / values (e.g. that query values are [1, 1, 1, 1]), etc.
By the way, should not we initialize it as SparseTextEmbedding(model_name="Qdrant/bm42-all-minilm-l6-v2-attentions")
?
If we want to initialize it as SparseTextEmbedding
, then we also need to overload methods like query_embed
in it.
Other than this, the PR looks ok
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
If we want to initialize it as SparseTextEmbedding, then we also need to overload methods like query_embed in it.
yeah, this is right
ee7a780
to
ad80e59
Compare
Intro new sparse model based on attention weights. Consider it as an extension of bm25 for short documents.