You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Add VisualQnA docker for both Gaudi and Xeon using TGI serving (opea-project#547)
* Add VisualQnA docker for both Gaudi and Xeon
Signed-off-by: lvliang-intel <liang1.lv@intel.com>
* [pre-commit.ci] auto fixes from pre-commit.com hooks
for more information, see https://pre-commit.ci
* update token length
Signed-off-by: lvliang-intel <liang1.lv@intel.com>
---------
Signed-off-by: lvliang-intel <liang1.lv@intel.com>
Co-authored-by: pre-commit-ci[bot] <66853113+pre-commit-ci[bot]@users.noreply.github.com>
Here are some explanation about the above parameters:
69
+
> Notice: Currently only the **Habana Driver 1.16.x** is supported for Gaudi.
72
70
73
-
-`--host`: the host of the gradio app
74
-
-`--port`: the port of the gradio app, by default 7860
75
-
-`--worker-addr`: the LLaVA service IP address. If you setup the service on a different machine, please replace `localhost` to the IP address of your Gaudi2 host machine
76
-
-`--lang`: Specify this parameter to use the Chinese interface. The default UI language is English and can be used without any additional parameter.
71
+
## Deploy VisualQnA on Xeon
77
72
78
-
SCRIPT USAGE NOTICE: By downloading and using any script file included with the associated software package (such as files with .bat, .cmd, or .JS extensions, Docker files, or any other type of file that, when executed, automatically downloads and/or installs files onto your system) (the “Script File”), it is your obligation to review the Script File to understand what files (e.g., other software, AI models, AI Datasets) the Script File will download to your system (“Downloaded Files”). Furthermore, by downloading and using the Downloaded Files, even if they are installed through a silent install, you agree to any and all terms and conditions associated with such files, including but not limited to, license terms, notices, or disclaimers.
73
+
Refer to the [Xeon Guide](./docker/xeon/README.md) for more instructions on building docker images from source.
74
+
75
+
Find the corresponding [compose.yaml](./docker/xeon/compose.yaml).
This document outlines the deployment process for a VisualQnA application utilizing the [GenAIComps](https://github.com/opea-project/GenAIComps.git) microservice pipeline on Intel Gaudi server. The steps include Docker image creation, container deployment via Docker Compose, and service execution to integrate microservices such as llm. We will publish the Docker images to Docker Hub, it will simplify the deployment process for this service.
4
+
5
+
## 🚀 Build Docker Images
6
+
7
+
First of all, you need to build Docker Images locally. This step can be ignored after the Docker images published to Docker hub.
To construct the Mega Service, we utilize the [GenAIComps](https://github.com/opea-project/GenAIComps.git) microservice pipeline within the `visuralqna.py` Python script. Build the MegaService Docker image using the command below:
To access the frontend, open the following URL in your browser: http://{host_ip}:5173. By default, the UI runs on port 5173 internally. If you prefer to use a different host port to access the frontend, you can modify the port mapping in the `compose.yaml` file as shown below:
0 commit comments