-
Notifications
You must be signed in to change notification settings - Fork 64
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Any examples or videos? #81
Comments
Hi @EmmaWebGH Thanks for your interest in the project. There are no videos, but a cli based demo using Google Colab. It is also linked in Readme - https://githubtocolab.com/snexus/llm-search/blob/main/notebooks/llmsearch_google_colab_demo.ipynb You can try it out on your documents and/or custom models. Please pay attention to the GPU limitation of the free Google Colab. If you think it suits your needs, the next step would be to install a local version with web based UI. Please follow the readme for that.
This is definitely on the radar, it would be good to decouple low-level model handling from RAG. |
Hello, |
Hi @amscosta When you open the notebook, you can click on the left pane, then right click -> create folder, as shown on the screenshot below: Name the folder |
Hi
b) When you prompted "ENTER QUESTION >> How to specify target branches in git?" |
Hi @amscosta
Yes, essentially it looks for answers in the provided docs, and should refuse if information isn't present there. The quality of the response depends on the underlying LLM model, which is configurable. Search for "RAG" architecture for more information. Pay attention that the question/answer interface provided in the demo is simplistic - for full experience you can install the package locally on your computer (provided your hardware is adequate, or you can use OpenAI's ChatGPT as the backend model), and use the web interface. |
Thank you for the clarification.
How big should be the sample_docs folder ?
For instance, how many files did you upload in the folder sample_docs for
the question : How to specify target branches in git?"
Em sex., 12 de jan. de 2024 às 09:48, Denis Lapchev <
***@***.***> escreveu:
… Hi @amscosta <https://github.com/amscosta>
sample_data folder is equivalent to sample_docs folder (you mentioned in
the colab)?
sampe_data is the default folder that Google Colab created for you. The
package expects sample_docs - which you need to create manually (as
explained in my previous reply), and upload your docs there (in one of the
supported formats). This is all configurable in the notebook, under the
cell Prepare configuration and download the model
When you prompted "ENTER QUESTION >> How to specify target branches in
git?"
The answer is generated by the LLM model with some more significant
"weights" in the contents of the "knowledge" embedded in the sample_docs
folder. Is that how it works?
Yes, essentially it looks for answers in the provided docs, and should
refuse if information isn't present there. The quality of the response
depends on the underlying LLM model, which is configurable. Search for
"RAG" architecture for more information.
Pay attention that the question/answer interface provided in the demo is
simplistic - for full experience you can install the package locally on
your computer (provided your hardware is adequate, or you can use OpenAI's
ChatGPT as the backend model), and use the web interface.
—
Reply to this email directly, view it on GitHub
<#81 (comment)>,
or unsubscribe
<https://github.com/notifications/unsubscribe-auth/AHTU55AK2KN5TFE7LPOFOUDYOEWJ3AVCNFSM6AAAAABBUKQO7KVHI2DSMVQWIX3LMV43OSLTON2WKQ3PNVWWK3TUHMYTQOBZGEYDGOJRGA>
.
You are receiving this because you were mentioned.Message ID:
***@***.***>
|
Hi, In the offline version - I use it with a 500MB-1 GB knowledge base (combined - pdf and markdown files). Don't think it will scale well beyond a few GBs. For the demo in Google Colab - 100 MB should be feasible. I used just a few markdown files to test that question you are mentioning. |
I can create a video if I have time. I think it would be useful and some aspects of this project require knowledge of llm parameters & how they work. Also I think some of the examples are somewhat outdated. |
That would be greatly appreciated @Hisma |
Hi team
Do you have any examples or know of any videos of people showing this?
I literally can't find a thing - but it sounds really good for RAG.
However, as you can imagine, searching web or Youtube for "llm search" is so generic, that the results contain anything and everything.
Even searching for "llm-search", nothing, just generic results for... llm and... search... and building search engines with llms...
I'd consider updating the project name.
Anyway, that said, this sounds like it has more/better RAG options than most other stuff I've been trying out. But I do really like to see demos of things too, before I spend time trying to get it to run.
As an aside, any plans to enable this to run via APIs so we can use it with Ollama or Oobabooga, as other tools can? This would be great for using all kinds of GPU-accelerated models.
Thanks!
The text was updated successfully, but these errors were encountered: