diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 15bafde5..21e1c623 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -50,7 +50,7 @@ application: - amd64 ports: - 8001 - image: quay.io/ai-lab/llamacpp-python:latest + image: quay.io/ai-lab/llamacpp_python:latest - name: streamlit-chat-app contextdir: . containerfile: app/Containerfile diff --git a/ailab-images.md b/ailab-images.md index cf462e23..5392f440 100644 --- a/ailab-images.md +++ b/ailab-images.md @@ -1,8 +1,8 @@ ## Images (x86_64, aarch64) currently built from GH Actions in this repository -- quay.io/ai-lab/llamacpp-python:latest -- quay.io/ai-lab/llamacpp-python-cuda:latest -- quay.io/ai-lab/llamacpp-python-vulkan:latest +- quay.io/ai-lab/llamacpp_python:latest +- quay.io/ai-lab/llamacpp_python_cuda:latest +- quay.io/ai-lab/llamacpp_python_vulkan:latest - quay.io/ai-lab/summarizer:latest - quay.io/ai-lab/chatbot:latest - quay.io/ai-lab/rag:latest diff --git a/model_servers/llamacpp_python/README.md b/model_servers/llamacpp_python/README.md index dc463831..6e69b27d 100644 --- a/model_servers/llamacpp_python/README.md +++ b/model_servers/llamacpp_python/README.md @@ -24,7 +24,7 @@ make -f Makefile build To pull the base model service image: ```bash -podman pull quay.io/ai-lab/llamacpp-python +podman pull quay.io/ai-lab/llamacpp_python ``` @@ -40,7 +40,7 @@ make -f Makefile build-cuda To pull the base model service image: ```bash -podman pull quay.io/ai-lab/llamacpp-python-cuda +podman pull quay.io/ai-lab/llamacpp_python_cuda ``` **IMPORTANT!** @@ -67,7 +67,7 @@ make -f Makefile build-vulkan To pull the base model service image: ```bash -podman pull quay.io/ai-lab/llamacpp-python-vulkan +podman pull quay.io/ai-lab/llamacpp_python_vulkan ``` diff --git a/model_servers/llamacpp_python/tests/conftest.py b/model_servers/llamacpp_python/tests/conftest.py index 7ceb2c6c..24cb9440 100644 --- a/model_servers/llamacpp_python/tests/conftest.py +++ b/model_servers/llamacpp_python/tests/conftest.py @@ -8,7 +8,7 @@ REGISTRY = os.environ['REGISTRY'] if not 'IMAGE_NAME' in os.environ: - IMAGE_NAME = 'containers/llamacpp-python:latest' + IMAGE_NAME = 'containers/llamacpp_python:latest' else: IMAGE_NAME = os.environ['IMAGE_NAME'] diff --git a/recipes/audio/audio_to_text/bootc/Containerfile b/recipes/audio/audio_to_text/bootc/Containerfile index b9626892..d53c26d1 100644 --- a/recipes/audio/audio_to_text/bootc/Containerfile +++ b/recipes/audio/audio_to_text/bootc/Containerfile @@ -16,7 +16,7 @@ RUN set -eu; mkdir -p /usr/ssh && \ ARG RECIPE=audio-to-text ARG MODEL_IMAGE=quay.io/ai-lab/mistral-7b-instruct:latest ARG APP_IMAGE=quay.io/ai-lab/${RECIPE}:latest -ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp-python:latest +ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp_python:latest ARG TARGETARCH # Add quadlet files to setup system to automatically run AI application on boot diff --git a/recipes/audio/audio_to_text/bootc/README.md b/recipes/audio/audio_to_text/bootc/README.md index e9cd2baa..d7512a92 100644 --- a/recipes/audio/audio_to_text/bootc/README.md +++ b/recipes/audio/audio_to_text/bootc/README.md @@ -32,7 +32,7 @@ podman build --build-arg "sshpubkey=$(cat ~/.ssh/id_rsa.pub)" \ # for GPU powered sample LLM application with llamacpp cuda model server podman build --build-arg "sshpubkey=$(cat ~/.ssh/id_rsa.pub)" \ - --build-arg "model-server-image="quay.io/ai-lab/llamacpp-python-cuda:latest" \ + --build-arg "model-server-image="quay.io/ai-lab/llamacpp_python_cuda:latest" \ --from \ --cap-add SYS_ADMIN \ --platform linux/amd64 \ diff --git a/recipes/common/Makefile.common b/recipes/common/Makefile.common index 31de210e..d16fc02a 100644 --- a/recipes/common/Makefile.common +++ b/recipes/common/Makefile.common @@ -5,7 +5,7 @@ IMAGE_NAME ?= $(REGISTRY_ORG)/${APP}:latest APP_IMAGE ?= $(REGISTRY)/$(IMAGE_NAME) CHROMADB_IMAGE ?= $(REGISTRY)/$(REGISTRY_ORG)/chromadb:latest MODEL_IMAGE ?= $(REGISTRY)/$(REGISTRY_ORG)/mistral-7b-instruct:latest -SERVER_IMAGE ?= $(REGISTRY)/$(REGISTRY_ORG)/llamacpp-python:latest +SERVER_IMAGE ?= $(REGISTRY)/$(REGISTRY_ORG)/llamacpp_python:latest SSH_PUBKEY ?= $(shell cat ${HOME}/.ssh/id_rsa.pub;) BOOTC_IMAGE ?= quay.io/$(REGISTRY_ORG)/${APP}-bootc:latest BOOTC_IMAGE_BUILDER ?= quay.io/centos-bootc/bootc-image-builder diff --git a/recipes/common/README.md b/recipes/common/README.md index 7316c02d..2316f698 100644 --- a/recipes/common/README.md +++ b/recipes/common/README.md @@ -29,7 +29,7 @@ used to override defaults for a variety of make targets. |CHROMADB_IMAGE | ChromaDB image to be used for application | `$(REGISTRY)/$(REGISTRY_ORG)/chromadb:latest` | |DISK_TYPE | Disk type to be created by BOOTC_IMAGE_BUILDER | `qcow2` (Options: ami, iso, vmdk, raw) | |MODEL_IMAGE | AI Model to be used by application | `$(REGISTRY)/$(REGISTRY_ORG)/mistral-7b-instruct:latest`| -|SERVER_IMAGE | AI Model Server Application | `$(REGISTRY)/$(REGISTRY_ORG)/llamacpp-python:latest` | +|SERVER_IMAGE | AI Model Server Application | `$(REGISTRY)/$(REGISTRY_ORG)/llamacpp_python:latest` | |SSH_PUBKEY | SSH Public key preloaded in bootc image. | `$(shell cat ${HOME}/.ssh/id_rsa.pub;)` | |FROM | Overrides first FROM instruction within Containerfile| `FROM` line defined in the Containerfile | |ARCH | Use alternate arch for image build | Current Arch | diff --git a/recipes/natural_language_processing/chatbot/ai-lab.yaml b/recipes/natural_language_processing/chatbot/ai-lab.yaml index 6003c2e3..17029c2b 100644 --- a/recipes/natural_language_processing/chatbot/ai-lab.yaml +++ b/recipes/natural_language_processing/chatbot/ai-lab.yaml @@ -15,7 +15,7 @@ application: - amd64 ports: - 8001 - image: quay.io/ai-lab/llamacpp-python:latest + image: quay.io/ai-lab/llamacppp_python:latest - name: streamlit-chat-app contextdir: app containerfile: Containerfile diff --git a/recipes/natural_language_processing/chatbot/bootc/Containerfile b/recipes/natural_language_processing/chatbot/bootc/Containerfile index 83fba11f..9907863e 100644 --- a/recipes/natural_language_processing/chatbot/bootc/Containerfile +++ b/recipes/natural_language_processing/chatbot/bootc/Containerfile @@ -16,7 +16,7 @@ RUN set -eu; mkdir -p /usr/ssh && \ ARG RECIPE=chatbot ARG MODEL_IMAGE=quay.io/ai-lab/mistral-7b-instruct:latest ARG APP_IMAGE=quay.io/ai-lab/${RECIPE}:latest -ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp-python:latest +ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp_python:latest ARG TARGETARCH # Add quadlet files to setup system to automatically run AI application on boot diff --git a/recipes/natural_language_processing/chatbot/bootc/README.md b/recipes/natural_language_processing/chatbot/bootc/README.md index d57e2d8b..e12ed92b 100644 --- a/recipes/natural_language_processing/chatbot/bootc/README.md +++ b/recipes/natural_language_processing/chatbot/bootc/README.md @@ -32,7 +32,7 @@ podman build --build-arg "sshpubkey=$(cat ~/.ssh/id_rsa.pub)" \ # for GPU powered sample LLM application with llamacpp cuda model server podman build --build-arg "sshpubkey=$(cat ~/.ssh/id_rsa.pub)" \ - --build-arg "model-server-image="quay.io/ai-lab/llamacpp-python-cuda:latest" \ + --build-arg "model-server-image="quay.io/ai-lab/llamacpp_python_cuda:latest" \ --from \ --cap-add SYS_ADMIN \ --platform linux/amd64 \ diff --git a/recipes/natural_language_processing/chatbot/provision/playbook.yml b/recipes/natural_language_processing/chatbot/provision/playbook.yml index 13e15124..366cabe3 100644 --- a/recipes/natural_language_processing/chatbot/provision/playbook.yml +++ b/recipes/natural_language_processing/chatbot/provision/playbook.yml @@ -33,7 +33,7 @@ - name: Run Model containers.podman.podman_container: name: llamacpp_python - image: ghcr.io/containers/llamacpp-python:latest + image: ghcr.io/containers/llamacpp_python:latest state: started interactive: true tty: true diff --git a/recipes/natural_language_processing/codegen/ai-lab.yaml b/recipes/natural_language_processing/codegen/ai-lab.yaml index 98e7dd14..dca3fb14 100644 --- a/recipes/natural_language_processing/codegen/ai-lab.yaml +++ b/recipes/natural_language_processing/codegen/ai-lab.yaml @@ -15,7 +15,7 @@ application: - amd64 ports: - 8001 - image: quay.io/ai-lab/llamacpp-python:latest + image: quay.io/ai-lab/llamacpp_python:latest - name: codegen-app contextdir: app containerfile: Containerfile diff --git a/recipes/natural_language_processing/codegen/bootc/Containerfile b/recipes/natural_language_processing/codegen/bootc/Containerfile index 9f0c1fae..3551e0f7 100644 --- a/recipes/natural_language_processing/codegen/bootc/Containerfile +++ b/recipes/natural_language_processing/codegen/bootc/Containerfile @@ -16,7 +16,7 @@ RUN set -eu; mkdir -p /usr/ssh && \ ARG RECIPE=codegen ARG MODEL_IMAGE=quay.io/ai-lab/mistral-7b-instruct:latest ARG APP_IMAGE=quay.io/ai-lab/${RECIPE}:latest -ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp-python:latest +ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp_python:latest ARG TARGETARCH # Add quadlet files to setup system to automatically run AI application on boot diff --git a/recipes/natural_language_processing/codegen/provision/playbook.yml b/recipes/natural_language_processing/codegen/provision/playbook.yml index 7ec5795d..15e061e5 100644 --- a/recipes/natural_language_processing/codegen/provision/playbook.yml +++ b/recipes/natural_language_processing/codegen/provision/playbook.yml @@ -33,7 +33,7 @@ - name: Run Model containers.podman.podman_container: name: llamacpp_python - image: ghcr.io/containers/llamacpp-python:latest + image: ghcr.io/containers/llamacpp_python:latest state: started interactive: true tty: true diff --git a/recipes/natural_language_processing/rag/ai-lab.yaml b/recipes/natural_language_processing/rag/ai-lab.yaml index fc6b5818..fbaf387f 100644 --- a/recipes/natural_language_processing/rag/ai-lab.yaml +++ b/recipes/natural_language_processing/rag/ai-lab.yaml @@ -15,7 +15,7 @@ application: - amd64 ports: - 8001 - image: quay.io/ai-lab/llamacpp-python:latest + image: quay.io/ai-lab/llamacpp_python:latest - name: chromadb-server contextdir: ../../../vector_dbs/chromadb containerfile: Containerfile diff --git a/recipes/natural_language_processing/rag/bootc/Containerfile b/recipes/natural_language_processing/rag/bootc/Containerfile index 37082bb2..0e7e7927 100644 --- a/recipes/natural_language_processing/rag/bootc/Containerfile +++ b/recipes/natural_language_processing/rag/bootc/Containerfile @@ -17,7 +17,7 @@ RUN set -eu; mkdir -p /usr/ssh && \ ARG RECIPE=rag ARG MODEL_IMAGE=quay.io/ai-lab/mistral-7b-instruct:latest ARG APP_IMAGE=quay.io/ai-lab/${RECIPE}:latest -ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp-python:latest +ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp_python:latest ARG CHROMADBImage=quay.io/ai-lab/chromadb ARG TARGETARCH diff --git a/recipes/natural_language_processing/rag/provision/playbook.yml b/recipes/natural_language_processing/rag/provision/playbook.yml index 2df7e964..2f127861 100644 --- a/recipes/natural_language_processing/rag/provision/playbook.yml +++ b/recipes/natural_language_processing/rag/provision/playbook.yml @@ -33,7 +33,7 @@ - name: Run Model containers.podman.podman_container: name: llamacpp_python - image: ghcr.io/containers/llamacpp-python:latest + image: ghcr.io/containers/llamacpp_python:latest state: started interactive: true tty: true diff --git a/recipes/natural_language_processing/summarizer/ai-lab.yaml b/recipes/natural_language_processing/summarizer/ai-lab.yaml index 26b76cb3..3ea24289 100644 --- a/recipes/natural_language_processing/summarizer/ai-lab.yaml +++ b/recipes/natural_language_processing/summarizer/ai-lab.yaml @@ -15,7 +15,7 @@ application: - amd64 ports: - 8001 - image: quay.io/ai-lab/llamacpp-python:latest + image: quay.io/ai-lab/llamacpp_python:latest - name: streamlit-summary-app contextdir: app containerfile: Containerfile diff --git a/recipes/natural_language_processing/summarizer/bootc/Containerfile b/recipes/natural_language_processing/summarizer/bootc/Containerfile index f38c335c..f44199fc 100644 --- a/recipes/natural_language_processing/summarizer/bootc/Containerfile +++ b/recipes/natural_language_processing/summarizer/bootc/Containerfile @@ -16,7 +16,7 @@ RUN set -eu; mkdir -p /usr/ssh && \ ARG RECIPE=summarizer ARG MODEL_IMAGE=quay.io/ai-lab/mistral-7b-instruct:latest ARG APP_IMAGE=quay.io/ai-lab/${RECIPE}:latest -ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp-python:latest +ARG SERVER_IMAGE=quay.io/ai-lab/llamacpp_python:latest ARG TARGETARCH # Add quadlet files to setup system to automatically run AI application on boot diff --git a/recipes/natural_language_processing/summarizer/provision/playbook.yml b/recipes/natural_language_processing/summarizer/provision/playbook.yml index 83bac04f..683d22e8 100644 --- a/recipes/natural_language_processing/summarizer/provision/playbook.yml +++ b/recipes/natural_language_processing/summarizer/provision/playbook.yml @@ -33,7 +33,7 @@ - name: Run Model containers.podman.podman_container: name: llamacpp_python - image: ghcr.io/containers/llamacpp-python:latest + image: ghcr.io/containers/llamacpp_python:latest state: started interactive: true tty: true