Skip to content
Snippets Groups Projects
Unverified Commit 4f651e38 authored by Henry Chen's avatar Henry Chen Committed by GitHub
Browse files

Pass `TRUNCATE_PROMPT_LENGTH` to deployment, switch to `ghcr.io` image (#63)

parent 2f03dee8
No related branches found
No related tags found
No related merge requests found
apiVersion: v2 apiVersion: v2
appVersion: 0.11.0 appVersion: 0.11.1
description: A Helm chart for ialacol description: A Helm chart for ialacol
name: ialacol name: ialacol
type: application type: application
version: 0.11.0 version: 0.11.1
...@@ -57,6 +57,8 @@ spec: ...@@ -57,6 +57,8 @@ spec:
value: {{ (.Values.deployment.env).GPU_LAYERS | quote }} value: {{ (.Values.deployment.env).GPU_LAYERS | quote }}
- name: MODE_TYPE - name: MODE_TYPE
value: {{ (.Values.deployment.env).MODE_TYPE | quote }} value: {{ (.Values.deployment.env).MODE_TYPE | quote }}
- name: TRUNCATE_PROMPT_LENGTH
value: {{ (.Values.deployment.env).TRUNCATE_PROMPT_LENGTH | quote }}
volumeMounts: volumeMounts:
- mountPath: /app/models - mountPath: /app/models
name: model name: model
......
replicas: 1 replicas: 1
deployment: deployment:
image: quay.io/chenhunghan/ialacol:latest image: ghcr.io/chenhunghan/ialacol:latest
# or use CUDA image `ghcr.io/chenhunghan/ialacol-cuda12:latest` # or use CUDA image `ghcr.io/chenhunghan/ialacol-cuda12:latest`
# env: # env:
# DEFAULT_MODEL_HG_REPO_ID: TheBloke/Llama-2-7B-Chat-GGML # DEFAULT_MODEL_HG_REPO_ID: TheBloke/Llama-2-7B-Chat-GGML
......
0% Loading or .
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment