You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The command below will launch the multi-agent system with the `DocIndexRetriever` as the retrieval tool for the Worker RAG agent.
113
-
114
-
```bash
115
-
docker compose -f $WORKDIR/GenAIExamples/DocIndexRetriever/docker_compose/intel/cpu/xeon/compose.yaml -f compose_openai.yaml up -d
116
-
```
117
-
118
-
#### Models on Remote Server
119
-
120
-
When models are deployed on a remote server with Intel® AI for Enterprise Inference, a base URL and an API key are required to access them. To run the Agent microservice on Gaudi while using models deployed on a remote server, add `compose_remote.yaml` to the `docker compose` command and set additional environment variables.
121
-
122
-
#### Notes
123
-
124
-
-`OPENAI_API_KEY` is already set in a previous step.
125
-
-`model` is used to overwrite the value set for this environment variable in `set_env.sh`.
126
-
-`LLM_ENDPOINT_URL` is the base URL given from the owner of the on-prem machine or cloud service provider. It will follow this format: "https://<DNS>". Here is an example: "https://api.inference.example.com".
0 commit comments