Skip to content

Offline LLM where the prompts are fine-tuned using IterDRAG

Notifications You must be signed in to change notification settings

vanlalpeka/Local_LLM

Repository files navigation

Local LLM with Streamlit UI

This container is based on ollama/ollama image to run LLM offline (local). The "stack" is deepseek-r1:1.5b -> langchain -> streamlit UI. To update the LLM model, change the model name in the ChatOllama() call.

# /src/app.py
# Currently using deepseek-r1:1.5b.
# Update the model as prefered

response = ChatOllama(model="deepseek-r1:1.5b",....)

IterDRAG technique is used for RAG.

The "</think>" tag is not printed on the Streamlit UI.

The prompt instructions are from Langchain's Ollama Deep Researcher.

Usage (Docker):

Pull the image from Docker hub and run.

docker pull vpeka/local_llm
&&
docker run -p 8501:8501 -p 11434:11434 --volume /host/models:/root/.ollama  vpeka/local_llm 

Usage (Linux/WSL): Clone this repo

Clone this repo, then run the following steps:

  1. Install Ollama
curl -fsSL https://ollama.com/install.sh | sh
  1. Install requirements
pip install -r requirements.txt
  1. Run Ollama in the background
ollama serve & sleep 5 && ollama run deepseek-r1:1.5b
  1. Run the streamlit app
streamlit run src/chat/app.py

About

Offline LLM where the prompts are fine-tuned using IterDRAG

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published