Skip to content

Commit

Permalink
Merge pull request #477 from h2oai/devops/chathurinda/readme-for-dock…
Browse files Browse the repository at this point in the history
…er-runtime

[DevOps] Update README for docker runtime image consumption
  • Loading branch information
ChathurindaRanasinghe authored Jul 17, 2023
2 parents c14002c + 3fbcdb1 commit 999632e
Showing 1 changed file with 38 additions and 0 deletions.
38 changes: 38 additions & 0 deletions docs/INSTALL-DOCKER.md
Original file line number Diff line number Diff line change
Expand Up @@ -47,6 +47,44 @@

4. Open `https://localhost:7860` in the browser


### Run h2oGPT using Docker runtime
1. Make sure Docker & Nvidia Containers are setup correctly by following step 1 in above instructions. Runtime docker images for h2oGPT are stored in the h2oai GCR vorvan container repository.

2. Specify the required model using `HF_MODEL` parameter.
All open-source datasets and models are posted on [🤗 H2O.ai's Hugging Face page](https://huggingface.co/h2oai/).
```bash
docker run \
--runtime=nvidia \
--shm-size=64g \
-e HF_MODEL=h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b \
-p 8888:8888 \
-p 7860:7860 \
--rm \
--init \
-v `pwd`/h2ogpt_env:/h2ogpt_env \
gcr.io/vorvan/h2oai/h2ogpt-runtime:61d6aea6fff3b1190aa42eee7fa10d6c
```
3. Navigate to http://localhost:7860/ & start using h2oGPT.
__Optional__
To run with a custom entrypoint, modify the local [`run-gpt.sh`](https://github.com/h2oai/h2ogpt/blob/76947c009a82d7a4a871548e68a60ce0a28b75d1/run-gpt.sh) & mount it.
```bash
docker run \
--runtime=nvidia \
--shm-size=64g \
-e HF_MODEL=h2oai/h2ogpt-gm-oasst1-en-2048-open-llama-7b \
-p 8888:8888 \
-p 7860:7860 \
--rm \
--init \
-v `pwd`/h2ogpt_env:/h2ogpt_env \
-v `pwd`/run-gpt.sh:/run-gpt.sh \
gcr.io/vorvan/h2oai/h2ogpt-runtime:61d6aea6fff3b1190aa42eee7fa10d6c
```
### Docker Compose Setup & Inference
1. (optional) Change desired model and weights under `environment` in the `docker-compose.yml`
Expand Down

0 comments on commit 999632e

Please sign in to comment.