Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

OSError: exception: access violation writing 0x0000000000000380 unable to query two querys at once the LLMchain(llm=LLamaCPP())) #2941

Closed
Revanth-guduru-balaji opened this issue Aug 31, 2023 · 4 comments
Labels

Comments

@Revanth-guduru-balaji
Copy link

I create a streamlit app to chat with Lamma 7B chat model.

When I open two tabs to chat and ask query it crashes becasue of the error OSError: exception: access violation writing 0x0000000000000380.

streaming was False
llm = LlamaCpp(
model_path="./models/llama-2-7b-chat.ggmlv3.q4_0.bin",
n_ctx=6000,
n_gpu_layers=512,
n_batch=512,
temperature = 0.9,
max_tokens = 64,
n_parts=1,
streaming=False
)

llm_chain = LLMChain(prompt=prompt, llm=llm)
llm_chain.run(st.session_state["chat_query"])

@ghost
Copy link

ghost commented Aug 31, 2023

open two tabs

Edit: In progress: #2813 (comment)

@Revanth-guduru-balaji
Copy link
Author

streamlit app

Not supported. README shows Supported platforms.

how to load model once in memory and use concurrently

@staviq
Copy link
Contributor

staviq commented Aug 31, 2023

how to load model once in memory and use concurrently

See #2813

@github-actions github-actions bot added the stale label Mar 21, 2024
Copy link
Contributor

github-actions bot commented Apr 5, 2024

This issue was closed because it has been inactive for 14 days since being marked as stale.

@github-actions github-actions bot closed this as completed Apr 5, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
Projects
None yet
Development

No branches or pull requests

2 participants