Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

no thinking , no answer .. only DONE #593

Open
RamazanGeven opened this issue Feb 6, 2025 · 3 comments
Open

no thinking , no answer .. only DONE #593

RamazanGeven opened this issue Feb 6, 2025 · 3 comments

Comments

@RamazanGeven
Copy link

Yesterday, i installed 2 lollms on 2 different machine and they worked well. today i wanted to delete and install again. After install , it works but no thinking, no answer just DONE

[INFO][2025-02-06 09:32:21] Selecting personality

  • Selecting active personality 0 ...[INFO][2025-02-06 09:32:21] ok
    Selected lollms
    [INFO][2025-02-06 09:32:21] Saving configuration
    INFO: ::1:52670 - "POST /select_personality HTTP/1.1" 200 OK
    INFO: ::1:52670 - "POST /get_config HTTP/1.1" 200 OK
    INFO: ::1:52671 - "POST /get_config HTTP/1.1" 200 OK
    Listing all personalitiesINFO: ::1:52671 - "GET /list_bindings HTTP/1.1" 200 OK
    INFO: ::1:52671 - "GET /get_available_models HTTP/1.1" 200 OK
    Listing modelsok
    INFO: ::1:52671 - "GET /list_models HTTP/1.1" 200 OK
    Getting active model
    ok
    INFO: ::1:52671 - "GET /get_active_model HTTP/1.1" 200 OK
    INFO: ::1:52671 - "POST /get_config HTTP/1.1" 200 OK
    INFO: ::1:52671 - "POST /get_personality_languages_list HTTP/1.1" 200 OK
    INFO: ::1:52671 - "POST /get_personality_language HTTP/1.1" 200 OK
    INFO: ::1:52671 - "GET /is_rt_on HTTP/1.1" 200 OK
    OK
    INFO: ::1:52670 - "GET /get_all_personalities HTTP/1.1" 200 OK
    Loading discussion for client QPkiRKA1wyFMagQbAAAN ... ok
    INFO: ::1:52671 - "GET /is_rt_on HTTP/1.1" 200 OK
    INFO: ::1:52670 - "POST /get_discussion_files_list HTTP/1.1" 200 OK
    INFO: ::1:52672 - "POST /get_discussion_files_list HTTP/1.1" 200 OK
    INFO: ::1:52673 - "GET /get_generation_status HTTP/1.1" 200 OK
    Starting message generation by lollms
    [INFO][2025-02-06 09:32:34] Text generation requested by client: QPkiRKA1wyFMagQbAAAN
    [INFO][2025-02-06 09:32:34] Started generation task
    [INFO][2025-02-06 09:32:34] Received message : who are you (3)
    [INFO][2025-02-06 09:32:34] prompt has 1192 tokens
    [INFO][2025-02-06 09:32:34] warmup for generating up to 2898 tokens
    Llama.generate: 1 prefix-match hit, remaining 415 prompt tokens to eval
    llama_perf_context_print: load time = 1079.49 ms
    llama_perf_context_print: prompt eval time = 0.00 ms / 415 tokens ( 0.00 ms per token, inf tokens per second)
    llama_perf_context_print: eval time = 0.00 ms / 1 runs ( 0.00 ms per token, inf tokens per second)
    llama_perf_context_print: total time = 7141.27 ms / 416 tokens
    [INFO][2025-02-06 09:32:42]
    Finished executing the generation

[INFO][2025-02-06 09:32:42] ## Done Generation ##

[INFO][2025-02-06 09:32:42] ╔══════════════════════════════════════════════════╗
[INFO][2025-02-06 09:32:42] ║ Done ║
[INFO][2025-02-06 09:32:42] ╚══════════════════════════════════════════════════╝
Llama.generate: 1 prefix-match hit, remaining 32 prompt tokens to eval
INFO: ::1:52686 - "POST /get_discussion_files_list HTTP/1.1" 200 OK
llama_perf_context_print: load time = 1079.49 ms
llama_perf_context_print: prompt eval time = 0.00 ms / 32 tokens ( 0.00 ms per token, inf tokens per second)
llama_perf_context_print: eval time = 0.00 ms / 1 runs ( 0.00 ms per token, inf tokens per second)
llama_perf_context_print: total time = 1120.33 ms / 33 tokens
[INFO][2025-02-06 09:32:43]

@RamazanGeven
Copy link
Author

Image

@ParisNeo
Copy link
Owner

ParisNeo commented Feb 6, 2025

Hi there,

Some models are rigid in terms of compliance to template, so I think with these, you would need to go to settings tab and in main configuration you can find a sub card named model template. Try selecting another template (not the lollms one) and it should work. If the model template does not exist in the list, make sure you take a look in the model card and manually set the template. Given the messages you are getting, nothing is wrong, it is just the model generating an end of sentence token.

Image

I see that you are using a fusion between O1 deepseekR1 and qwen. This can happen as fine tuning may render the output model very template sensitive.

I will add an option to automatically extract the template format from the model itself, it became possible with the new versions of the binding.

Sorry for the inconveniance. I'm building this for free, alone and without any funds or external help.

@RamazanGeven
Copy link
Author

Thanks your great job. And about "Sorry for the inconveniance. I'm building this for free, alone and without any funds or external help." i know what it means very well i am a delphi developer more then 30 years.. was alone any fund and no external help.. So, i would like to help you if i can but after 1-2 months.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants