Skip to content

Commit 9b71d65

Browse files
authored
upgrade ghcr.io/huggingface/tei-gaudi:1.5.0 into ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 (#2119)
Signed-off-by: chensuyue <[email protected]>
1 parent fe255f2 commit 9b71d65

File tree

11 files changed

+16
-16
lines changed

11 files changed

+16
-16
lines changed

ChatQnA/docker_compose/intel/hpu/gaudi/README.md

Lines changed: 6 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -92,7 +92,7 @@ f090fe262c74 opea/chatqna-ui:latest
9292
ec97d7651c96 opea/chatqna:latest "python chatqna.py" 2 minutes ago Up 2 minutes 0.0.0.0:8888->8888/tcp, :::8888->8888/tcp chatqna-gaudi-backend-server
9393
a61fb7dc4fae opea/dataprep:latest "sh -c 'python $( [ …" 2 minutes ago Up 2 minutes 0.0.0.0:6007->5000/tcp, [::]:6007->5000/tcp dataprep-redis-server
9494
d560c232b120 opea/retriever:latest "python opea_retriev…" 2 minutes ago Up 2 minutes 0.0.0.0:7000->7000/tcp, :::7000->7000/tcp retriever-redis-server
95-
a1d7ca2d3787 ghcr.io/huggingface/tei-gaudi:1.5.0 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8808->80/tcp, [::]:8808->80/tcp tei-reranking-gaudi-server
95+
a1d7ca2d3787 ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8808->80/tcp, [::]:8808->80/tcp tei-reranking-gaudi-server
9696
9a9f3fd4fd4c opea/vllm-gaudi:latest "python3 -m vllm.ent…" 2 minutes ago Exited (1) 2 minutes ago vllm-gaudi-server
9797
1ab9bbdf5182 redis/redis-stack:7.2.0-v9 "/entrypoint.sh" 2 minutes ago Up 2 minutes 0.0.0.0:6379->6379/tcp, :::6379->6379/tcp, 0.0.0.0:8001->8001/tcp, :::8001->8001/tcp redis-vector-db
9898
9ee0789d819e ghcr.io/huggingface/text-embeddings-inference:cpu-1.7 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8090->80/tcp, [::]:8090->80/tcp tei-embedding-gaudi-server
@@ -150,7 +150,7 @@ The default deployment utilizes Gaudi devices primarily for the `vllm-service`,
150150
| dataprep-redis-service | opea/dataprep:latest | No |
151151
| tei-embedding-service | ghcr.io/huggingface/text-embeddings-inference:cpu-1.7 | No |
152152
| retriever | opea/retriever:latest | No |
153-
| tei-reranking-service | ghcr.io/huggingface/tei-gaudi:1.5.0 | 1 card |
153+
| tei-reranking-service | ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 | 1 card |
154154
| vllm-service | opea/vllm-gaudi:latest | Configurable |
155155
| chatqna-gaudi-backend-server | opea/chatqna:latest | No |
156156
| chatqna-gaudi-ui-server | opea/chatqna-ui:latest | No |
@@ -166,7 +166,7 @@ The TGI (Text Generation Inference) deployment and the default deployment differ
166166
| dataprep-redis-service | opea/dataprep:latest | No |
167167
| tei-embedding-service | ghcr.io/huggingface/text-embeddings-inference:cpu-1.7 | No |
168168
| retriever | opea/retriever:latest | No |
169-
| tei-reranking-service | ghcr.io/huggingface/tei-gaudi:1.5.0 | 1 card |
169+
| tei-reranking-service | ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 | 1 card |
170170
| **tgi-service** | ghcr.io/huggingface/tgi-gaudi:2.3.1 | Configurable |
171171
| chatqna-gaudi-backend-server | opea/chatqna:latest | No |
172172
| chatqna-gaudi-ui-server | opea/chatqna-ui:latest | No |
@@ -186,7 +186,7 @@ The TGI (Text Generation Inference) deployment and the default deployment differ
186186
| dataprep-redis-service | opea/dataprep:latest | No |
187187
| tei-embedding-service | ghcr.io/huggingface/text-embeddings-inference:cpu-1.7 | No |
188188
| retriever | opea/retriever:latest | No |
189-
| tei-reranking-service | ghcr.io/huggingface/tei-gaudi:1.5.0 | 1 card |
189+
| tei-reranking-service | ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 | 1 card |
190190
| vllm-service | opea/vllm-gaudi:latest | Configurable |
191191
| **llm-faqgen** | **opea/llm-faqgen:latest** | No |
192192
| chatqna-gaudi-backend-server | opea/chatqna:latest | No |
@@ -224,7 +224,7 @@ The _compose_guardrails.yaml_ Docker Compose file introduces enhancements over t
224224
| _guardrails_ | opea/guardrails:latest | No | No |
225225
| tei-embedding-service | ghcr.io/huggingface/text-embeddings-inference:cpu-1.7 | No | No |
226226
| retriever | opea/retriever:latest | No | No |
227-
| tei-reranking-service | ghcr.io/huggingface/tei-gaudi:1.5.0 | 1 card | No |
227+
| tei-reranking-service | ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 | 1 card | No |
228228
| vllm-service | opea/vllm-gaudi:latest | Configurable | Yes |
229229
| chatqna-gaudi-backend-server | opea/chatqna:latest | No | No |
230230
| chatqna-gaudi-ui-server | opea/chatqna-ui:latest | No | No |
@@ -260,7 +260,7 @@ The table provides a comprehensive overview of the ChatQnA services utilized acr
260260
| dataprep-redis-service | opea/dataprep:latest | No | Prepares data and interacts with the Redis database. |
261261
| tei-embedding-service | ghcr.io/huggingface/text-embeddings-inference:cpu-1.7 | No | Provides text embedding services, often using Hugging Face models. |
262262
| retriever | opea/retriever:latest | No | Retrieves data from the Redis database and interacts with embedding services. |
263-
| tei-reranking-service | ghcr.io/huggingface/tei-gaudi:1.5.0 | Yes | Reranks text embeddings, typically using Gaudi hardware for enhanced performance. |
263+
| tei-reranking-service | ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 | Yes | Reranks text embeddings, typically using Gaudi hardware for enhanced performance. |
264264
| vllm-service | opea/vllm-gaudi:latest | No | Handles large language model (LLM) tasks, utilizing Gaudi hardware. |
265265
| tgi-service | ghcr.io/huggingface/tgi-gaudi:2.3.1 | Yes | Specific to the TGI deployment, focuses on text generation inference using Gaudi hardware. |
266266
| tgi-guardrails-service | ghcr.io/huggingface/tgi-gaudi:2.3.1 | Yes | Provides guardrails functionality, ensuring safe operations within defined limits. |

ChatQnA/docker_compose/intel/hpu/gaudi/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -70,7 +70,7 @@ services:
7070
HF_TOKEN: ${HF_TOKEN}
7171
restart: unless-stopped
7272
tei-reranking-service:
73-
image: ghcr.io/huggingface/tei-gaudi:1.5.0
73+
image: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
7474
container_name: tei-reranking-gaudi-server
7575
ports:
7676
- "8808:80"

ChatQnA/docker_compose/intel/hpu/gaudi/compose_faqgen.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ services:
6464
HF_TOKEN: ${HF_TOKEN}
6565
restart: unless-stopped
6666
tei-reranking-service:
67-
image: ghcr.io/huggingface/tei-gaudi:1.5.0
67+
image: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
6868
container_name: tei-reranking-gaudi-server
6969
ports:
7070
- "8808:80"

ChatQnA/docker_compose/intel/hpu/gaudi/compose_faqgen_tgi.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -64,7 +64,7 @@ services:
6464
HF_TOKEN: ${HF_TOKEN}
6565
restart: unless-stopped
6666
tei-reranking-service:
67-
image: ghcr.io/huggingface/tei-gaudi:1.5.0
67+
image: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
6868
container_name: tei-reranking-gaudi-server
6969
ports:
7070
- "8808:80"

ChatQnA/docker_compose/intel/hpu/gaudi/compose_guardrails.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -109,7 +109,7 @@ services:
109109
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
110110
restart: unless-stopped
111111
tei-reranking-service:
112-
image: ghcr.io/huggingface/tei-gaudi:1.5.0
112+
image: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
113113
container_name: tei-reranking-gaudi-server
114114
ports:
115115
- "8808:80"

ChatQnA/docker_compose/intel/hpu/gaudi/compose_tgi.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -65,7 +65,7 @@ services:
6565
RETRIEVER_COMPONENT_NAME: "OPEA_RETRIEVER_REDIS"
6666
restart: unless-stopped
6767
tei-reranking-service:
68-
image: ghcr.io/huggingface/tei-gaudi:1.5.0
68+
image: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
6969
container_name: tei-reranking-gaudi-server
7070
ports:
7171
- "8808:80"

ChatQnA/docker_compose/intel/hpu/gaudi/how_to_validate_service.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -48,7 +48,7 @@ f810f3b4d329 opea/embedding:latest "python embed
4848
2fa17d84605f opea/dataprep:latest "python prepare_doc_…" 2 minutes ago Up 2 minutes 0.0.0.0:6007->6007/tcp, :::6007->5000/tcp dataprep-redis-server
4949
69e1fb59e92c opea/retriever:latest "/home/user/comps/re…" 2 minutes ago Up 2 minutes 0.0.0.0:7000->7000/tcp, :::7000->7000/tcp retriever-redis-server
5050
313b9d14928a opea/reranking-tei:latest "python reranking_te…" 2 minutes ago Up 2 minutes 0.0.0.0:8000->8000/tcp, :::8000->8000/tcp reranking-tei-gaudi-server
51-
174bd43fa6b5 ghcr.io/huggingface/tei-gaudi:1.5.0 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8090->80/tcp, :::8090->80/tcp tei-embedding-gaudi-server
51+
174bd43fa6b5 ghcr.io/huggingface/text-embeddings-inference:hpu-1.7 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8090->80/tcp, :::8090->80/tcp tei-embedding-gaudi-server
5252
05c40b636239 ghcr.io/huggingface/tgi-gaudi:2.3.1 "text-generation-lau…" 2 minutes ago Exited (1) About a minute ago tgi-gaudi-server
5353
74084469aa33 redis/redis-stack:7.2.0-v9 "/entrypoint.sh" 2 minutes ago Up 2 minutes 0.0.0.0:6379->6379/tcp, :::6379->6379/tcp, 0.0.0.0:8001->8001/tcp, :::8001->8001/tcp redis-vector-db
5454
88399dbc9e43 ghcr.io/huggingface/text-embeddings-inference:cpu-1.7 "text-embeddings-rou…" 2 minutes ago Up 2 minutes 0.0.0.0:8808->80/tcp, :::8808->80/tcp tei-reranking-gaudi-server

ChatQnA/kubernetes/gmc/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -24,7 +24,7 @@ The ChatQnA uses the below prebuilt images if you choose a Xeon deployment
2424
Should you desire to use the Gaudi accelerator, two alternate images are used for the embedding and llm services.
2525
For Gaudi:
2626

27-
tei-embedding-service: ghcr.io/huggingface/tei-gaudi:1.5.0
27+
tei-embedding-service: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
2828
tgi-service: gghcr.io/huggingface/tgi-gaudi:2.3.1
2929

3030

DocIndexRetriever/docker_compose/intel/hpu/gaudi/compose.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ services:
3636
retries: 10
3737
restart: unless-stopped
3838
tei-embedding-service:
39-
image: ghcr.io/huggingface/tei-gaudi:1.5.0
39+
image: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
4040
entrypoint: /bin/sh -c "apt-get update && apt-get install -y curl && text-embeddings-router --json-output --model-id ${EMBEDDING_MODEL_ID} --auto-truncate"
4141
container_name: tei-embedding-gaudi-server
4242
ports:

DocIndexRetriever/docker_compose/intel/hpu/gaudi/compose_milvus.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -95,7 +95,7 @@ services:
9595
condition: service_healthy
9696

9797
tei-embedding-service:
98-
image: ghcr.io/huggingface/tei-gaudi:1.5.0
98+
image: ghcr.io/huggingface/text-embeddings-inference:hpu-1.7
9999
entrypoint: /bin/sh -c "apt-get update && apt-get install -y curl && text-embeddings-router --json-output --model-id ${EMBEDDING_MODEL_ID} --auto-truncate"
100100
container_name: tei-embedding-gaudi-server
101101
ports:

0 commit comments

Comments
 (0)